|
3099 | 3099 | "forge/test/models/pytorch/vision/pointpillars/test_pointpillars.py::test_pointpillars": 0.598, |
3100 | 3100 | "forge/test/models/pytorch/vision/segformer/test_segformer.py::test_segformer_semantic_segmentation_pytorch[nvidia/segformer-b4-finetuned-ade-512-512]": 619.508, |
3101 | 3101 | "forge/test/models/pytorch/vision/unet/test_unet.py::test_unet_qubvel_pytorch": 50.701, |
3102 | | - "forge/test/models/pytorch/vision/unet/test_unet.py::test_unet_torchhub_pytorch": 101.852, |
| 3102 | + "forge/test/models/pytorch/vision/unet/test_unet.py::test_unet_torchhub_pytorch": 203.704, |
3103 | 3103 | "forge/test/models/pytorch/vision/yolo/test_yolo_world.py::test_yolo_world_inference": 45.038, |
3104 | 3104 | "forge/test/models/paddlepaddle/text/glm/test_glm.py::test_glm[THUDM/glm-515m]": 565.3, |
3105 | 3105 | "forge/test/models/paddlepaddle/text/t5/test_t5.py::test_t5_conditional_generation[t5-small]": 21.172, |
|
21121 | 21121 | "forge/test/models/pytorch/vision/yolo/test_yolo_v6.py::test_yolo_v6_pytorch[yolov6l]": 280.243, |
21122 | 21122 | "forge/test/models/pytorch/vision/yolo/test_yolox.py::test_yolox_pytorch[yolox_nano]": 119.101, |
21123 | 21123 | "forge/test/models/paddlepaddle/text/bert/test_bert.py::test_bert_question_answering[bert-base-uncased-input0]": 102.993, |
21124 | | - "forge/test/models/paddlepaddle/vision/resnet/test_resnet.py::test_resnet_pd[resnet34]": 40.791, |
| 21124 | + "forge/test/models/paddlepaddle/vision/resnet/test_resnet.py::test_resnet_pd[resnet34]": 81.582, |
21125 | 21125 | "forge/test/models/paddlepaddle/vision/resnet/test_resnet.py::test_resnet_pd[resnet152]": 138.265, |
21126 | 21126 | "forge/test/models/onnx/text/bert/test_bert.py::test_bert_masked_lm_onnx[17-bert-base-uncased]": 123.335, |
21127 | 21127 | "forge/test/models/onnx/vision/deit/test_deit_onnx.py::test_deit_onnx[facebook/deit-small-patch16-224]": 117.438, |
|
50200 | 50200 | "forge/test/models/pytorch/vision/segformer/test_segformer.py::test_segformer_semantic_segmentation_pytorch[nvidia/segformer-b3-finetuned-ade-512-512]": 708.26, |
50201 | 50201 | "forge/test/models/pytorch/vision/unet/test_unet.py::test_unet_osmr_cityscape_pytorch": 31.535, |
50202 | 50202 | "forge/test/models/pytorch/vision/yolo/test_yolo_v5.py::test_yolov5_320x320[yolov5x]": 54.41, |
50203 | | - "forge/test/models/paddlepaddle/vision/googlenet/test_googlenet.py::test_googlenet": 66.837, |
| 50203 | + "forge/test/models/paddlepaddle/vision/googlenet/test_googlenet.py::test_googlenet": 133.674, |
50204 | 50204 | "forge/test/models/onnx/vision/segformer/test_segformer.py::test_segformer_semantic_segmentation_onnx[nvidia/segformer-b1-finetuned-ade-512-512]": 116.086, |
50205 | 50205 | "forge/test/models/onnx/vision/segformer/test_segformer.py::test_segformer_semantic_segmentation_onnx[nvidia/segformer-b2-finetuned-ade-512-512]": 369.02, |
50206 | | - "forge/test/models/onnx/audio/test_whisper_onnx.py::test_whisper_onnx[openai/whisper-medium]": 6501.96, |
| 50206 | + "forge/test/models/onnx/audio/test_whisper_onnx.py::test_whisper_onnx[openai/whisper-medium]": 8127.45, |
50207 | 50207 | "forge/test/models/onnx/vision/perceiverio/test_perceiverio_onnx.py::test_perceiverio_for_image_classification_onnx[deepmind/vision-perceiver-conv]": 3213.3 |
50208 | 50208 | } |
0 commit comments