|
18548 | 18548 | "forge/test/models/pytorch/vision/yolo/test_yolo_v5.py::test_yolov5_640x640[yolov5s]": 88.938, |
18549 | 18549 | "forge/test/models/pytorch/vision/yolo/test_yolo_v5.py::test_yolov5_480x480[yolov5x]": 135.222, |
18550 | 18550 | "forge/test/models/paddlepaddle/text/bert/test_bert.py::test_bert_maskedlm[bert-base-uncased-input0]": 135.205, |
18551 | | - "forge/test/models/paddlepaddle/text/bert/test_bert.py::test_bert_maskedlm[uer/chinese-roberta-base-input2]": 278.9, |
| 18551 | + "forge/test/models/paddlepaddle/text/bert/test_bert.py::test_bert_maskedlm[uer/chinese-roberta-base-input2]": 478.9, |
18552 | 18552 | "forge/test/models/paddlepaddle/vision/densenet/test_densenet.py::test_densenet_pd[densenet121]": 260.748, |
18553 | 18553 | "forge/test/models/onnx/text/phi1/test_phi1_onnx.py::test_phi1_clm_onnx[microsoft/phi-1_5]": 0.001, |
18554 | 18554 | "forge/test/models/onnx/vision/deit/test_deit_onnx.py::test_deit_onnx[facebook/deit-base-patch16-224]": 121.457, |
|
21036 | 21036 | "forge/test/models/pytorch/text/mistral/test_mistral.py::test_mistral_v0_3[mistralai/Mistral-7B-Instruct-v0.3]": 0.884, |
21037 | 21037 | "forge/test/models/pytorch/text/mistral/test_mistral.py::test_mistral_Nemo[mistralai/Mistral-Nemo-Instruct-2407]": 0.929, |
21038 | 21038 | "forge/test/models/pytorch/text/opt/test_opt.py::test_opt_causal_lm[facebook/opt-1.3b]": 441.296, |
21039 | | - "forge/test/models/pytorch/text/opt/test_opt.py::test_opt_qa[facebook/opt-1.3b]": 159.33, |
| 21039 | + "forge/test/models/pytorch/text/opt/test_opt.py::test_opt_qa[facebook/opt-1.3b]": 239.33, |
21040 | 21040 | "forge/test/models/pytorch/text/phi3/test_phi3.py::test_phi3_causal_lm[microsoft/phi-3-mini-128k-instruct]": 1.335, |
21041 | 21041 | "forge/test/models/pytorch/text/phi3/test_phi3_5.py::test_phi3_5_causal_lm[microsoft/Phi-3.5-mini-instruct]": 0.489, |
21042 | 21042 | "forge/test/models/pytorch/text/phi4/test_phi4.py::test_phi_4_causal_lm_pytorch[microsoft/phi-4]": 0.489, |
|
21090 | 21090 | "forge/test/models/paddlepaddle/multimodal/paddleocr/test_paddleocr_rec.py::test_paddleocr_rec[v4_rec_en-https://paddleocr.bj.bcebos.com/PP-OCRv4/english/en_PP-OCRv4_rec_infer.tar]": 112.871, |
21091 | 21091 | "forge/test/models/paddlepaddle/multimodal/paddleocr/test_paddleocr_rec.py::test_paddleocr_rec[v0_rec_ch-https://paddleocr.bj.bcebos.com/dygraph_v2.0/ch/ch_ppocr_mobile_v2.0_rec_infer.tar]": 158.282, |
21092 | 21092 | "forge/test/models/paddlepaddle/vision/mobilenet/test_mobilenet_v2.py::test_mobilenetv2_basic": 59.963, |
21093 | | - "forge/test/models/paddlepaddle/vision/resnet/test_resnet.py::test_resnet_pd[resnet101]": 76.547, |
| 21093 | + "forge/test/models/paddlepaddle/vision/resnet/test_resnet.py::test_resnet_pd[resnet101]": 176.547, |
21094 | 21094 | "forge/test/models/onnx/vision/dla/test_dla.py::test_dla_onnx[dla46x_c]": 46.858, |
21095 | 21095 | "forge/test/models/onnx/vision/dla/test_dla.py::test_dla_onnx[dla60x]": 55.109, |
21096 | 21096 | "forge/test/models/onnx/vision/hrnet/test_hrnet_onnx.py::test_hrnet_onnx[hrnetv2_w64]": 370.458536, |
|
23500 | 23500 | "forge/test/models/paddlepaddle/multimodal/blip/test_blip.py::test_blip_text[Salesforce/blip-image-captioning-base]": 102.226, |
23501 | 23501 | "forge/test/models/paddlepaddle/text/albert/test_albert.py::test_albert_maskedlm[input0-albert-chinese-tiny]": 37.596, |
23502 | 23502 | "forge/test/models/paddlepaddle/text/bert/test_bert.py::test_bert_question_answering[cl-tohoku/bert-base-japanese-input1]": 157.885, |
23503 | | - "forge/test/models/paddlepaddle/text/ernie/test_ernie.py::test_ernie_for_sequence_classification[ernie-1.0]": 91.608, |
| 23503 | + "forge/test/models/paddlepaddle/text/ernie/test_ernie.py::test_ernie_for_sequence_classification[ernie-1.0]": 101.608, |
23504 | 23504 | "forge/test/models/paddlepaddle/text/roberta/test_roberta.py::test_roberta_sequence_classification[hfl/rbt4]": 102.64, |
23505 | 23505 | "forge/test/models/onnx/text/phi1/test_phi1_onnx.py::test_phi1_clm_onnx[microsoft/phi-1]": 0.001, |
23506 | 23506 | "forge/test/models/onnx/vision/alexnet/test_alexnet_onnx.py::test_alexnet_onnx": 21.733, |
|
23526 | 23526 | "forge/test/models/pytorch/vision/transfuser/test_transfuser.py::test_transfuser": 1.035, |
23527 | 23527 | "forge/test/models/pytorch/vision/yolo/test_yolo_v4.py::test_yolov4_tiny": 0.859, |
23528 | 23528 | "forge/test/models/paddlepaddle/multimodal/blip/test_blip.py::test_blip_vision[Salesforce/blip-image-captioning-base]": 15.629, |
23529 | | - "forge/test/models/paddlepaddle/multimodal/blip/test_blip.py::test_blip[Salesforce/blip-image-captioning-base]": 457.72, |
| 23529 | + "forge/test/models/paddlepaddle/multimodal/blip/test_blip.py::test_blip[Salesforce/blip-image-captioning-base]": 477.72, |
23530 | 23530 | "forge/test/models/paddlepaddle/text/glm/test_glm.py::test_glm[THUDM/glm-large-chinese]": 163.16, |
23531 | 23531 | "forge/test/models/pytorch/atomic/hippynn/test_hippynn.py::test_hippynn": 1.047, |
23532 | 23532 | "forge/test/models/pytorch/text/flux/test_flux.py::test_flux[black-forest-labs/FLUX.1-schnell]": 0.549, |
|
23585 | 23585 | "forge/test/models/pytorch/vision/ssd300_vgg16/test_ssd300_vgg16.py::test_ssd300_vgg16[ssd300_vgg16]": 3.882, |
23586 | 23586 | "forge/test/models/pytorch/vision/swin/test_swin.py::test_swin_v2_tiny_masked[microsoft/swinv2-tiny-patch4-window8-256]": 165.3, |
23587 | 23587 | "forge/test/models/paddlepaddle/multimodal/clip/test_clip.py::test_clip_text[openai/clip-vit-base-patch16]": 69.728, |
23588 | | - "forge/test/models/paddlepaddle/multimodal/clip/test_clip.py::test_clip_vision[openai/clip-vit-base-patch16]": 10.65, |
| 23588 | + "forge/test/models/paddlepaddle/multimodal/clip/test_clip.py::test_clip_vision[openai/clip-vit-base-patch16]": 200.65, |
23589 | 23589 | "forge/test/models/paddlepaddle/multimodal/clip/test_clip.py::test_clip[openai/clip-vit-base-patch16]": 18.313, |
23590 | 23590 | "forge/test/models/onnx/vision/efficientnet/test_efficientnet.py::test_efficientnet_onnx[efficientnet_b3]": 268.184081, |
23591 | 23591 | "forge/test/models/onnx/vision/xception/test_xception_onnx.py::test_xception_onnx[xception71.tf_in1k]": 103.353, |
@@ -50249,10 +50249,10 @@ |
50249 | 50249 | "forge/test/models/pytorch/text/bert/test_bert.py::test_bert_token_classification_pytorch[dbmdz/bert-large-cased-finetuned-conll03-english]": 99.12639, |
50250 | 50250 | "forge/test/models/pytorch/vision/segformer/test_segformer.py::test_segformer_image_classification_pytorch[mit_b5]": 588.587695, |
50251 | 50251 | "forge/test/models/pytorch/text/llama/test_llama3.py::test_llama3_causal_lm_pytorch[llama_3_8b_instruct]": 0.112146, |
50252 | | - "forge/test/models/pytorch/text/qwen/test_qwen_v3.py::test_qwen3_clm_pytorch[1_7b]": 446.14, |
| 50252 | + "forge/test/models/pytorch/text/qwen/test_qwen_v3.py::test_qwen3_clm_pytorch[1_7b]": 466.14, |
50253 | 50253 | "forge/test/models/pytorch/text/qwen/test_qwen_v2_5_coder.py::test_qwen_coder_clm_pytorch[1_5b]": 307.70, |
50254 | 50254 | "forge/test/models/onnx/vision/perceiverio/test_perceiverio_onnx.py::test_perceiverio_for_image_classification_onnx[deepmind/vision-perceiver-learned]": 854.83, |
50255 | 50255 | "forge/test/models/pytorch/text/phi1/test_phi1_5.py::test_phi1_5_sequence_classification_pytorch[microsoft/phi-1_5]": 220.04, |
50256 | | - "forge/test/models/pytorch/text/phi1/test_phi1_5.py::test_phi1_5_token_classification_pytorch[microsoft/phi-1_5]": 367.22, |
50257 | | - "forge/test/models/pytorch/vision/yolo/test_yolo_v8.py::test_yolov8[yolov8x]": 335.56 |
| 50256 | + "forge/test/models/pytorch/text/phi1/test_phi1_5.py::test_phi1_5_token_classification_pytorch[microsoft/phi-1_5]": 387.22, |
| 50257 | + "forge/test/models/pytorch/vision/yolo/test_yolo_v8.py::test_yolov8[yolov8x]": 355.56 |
50258 | 50258 | } |
0 commit comments