99 ModelFormat ,
1010 ModelVersion ,
1111 Protocols ,
12- RuntimeTemplates ,
12+ RunTimeConfigs ,
1313)
1414from utilities .inference_utils import Inference
1515from utilities .manifests .onnx import ONNX_INFERENCE_CONFIG
2424
2525@pytest .mark .serverless
2626@pytest .mark .parametrize (
27- "unprivileged_model_namespace, serving_runtime_from_template, s3_models_inference_service " ,
27+ "unprivileged_model_namespace, ovms_kserve_serving_runtime, ovms_kserve_inference_service " ,
2828 [
2929 pytest .param (
3030 {"name" : "kserve-ovms-metrics" },
31- {
32- "name" : f"{ Protocols .HTTP } -{ ModelFormat .ONNX } " ,
33- "template-name" : RuntimeTemplates .OVMS_KSERVE ,
34- "multi-model" : False ,
35- "enable-http" : True ,
36- },
31+ RunTimeConfigs .ONNX_OPSET13_RUNTIME_CONFIG ,
3732 {
3833 "name" : f"{ Protocols .HTTP } -{ ModelFormat .ONNX } " ,
3934 "deployment-mode" : KServeDeploymentType .SERVERLESS ,
4742class TestModelMetrics :
4843 @pytest .mark .smoke
4944 @pytest .mark .polarion ("ODS-2555" )
50- def test_model_metrics_num_success_requests (self , s3_models_inference_service , prometheus ):
45+ def test_model_metrics_num_success_requests (self , ovms_kserve_inference_service , prometheus ):
5146 """Verify number of successful model requests in OpenShift monitoring system (UserWorkloadMonitoring) metrics"""
5247 verify_inference_response (
53- inference_service = s3_models_inference_service ,
48+ inference_service = ovms_kserve_inference_service ,
5449 inference_config = ONNX_INFERENCE_CONFIG ,
5550 inference_type = Inference .INFER ,
5651 protocol = Protocols .HTTPS ,
@@ -64,12 +59,12 @@ def test_model_metrics_num_success_requests(self, s3_models_inference_service, p
6459
6560 @pytest .mark .smoke
6661 @pytest .mark .polarion ("ODS-2555" )
67- def test_model_metrics_num_total_requests (self , s3_models_inference_service , prometheus ):
62+ def test_model_metrics_num_total_requests (self , ovms_kserve_inference_service , prometheus ):
6863 """Verify number of total model requests in OpenShift monitoring system (UserWorkloadMonitoring) metrics"""
6964 total_runs = 5
7065
7166 run_inference_multiple_times (
72- isvc = s3_models_inference_service ,
67+ isvc = ovms_kserve_inference_service ,
7368 inference_config = ONNX_INFERENCE_CONFIG ,
7469 inference_type = Inference .INFER ,
7570 protocol = Protocols .HTTPS ,
@@ -84,9 +79,9 @@ def test_model_metrics_num_total_requests(self, s3_models_inference_service, pro
8479
8580 @pytest .mark .smoke
8681 @pytest .mark .polarion ("ODS-2555" )
87- def test_model_metrics_cpu_utilization (self , s3_models_inference_service , prometheus ):
82+ def test_model_metrics_cpu_utilization (self , ovms_kserve_inference_service , prometheus ):
8883 """Verify CPU utilization data in OpenShift monitoring system (UserWorkloadMonitoring) metrics"""
8984 assert get_metrics_value (
9085 prometheus = prometheus ,
91- metrics_query = f"pod:container_cpu_usage:sum{{namespace='{ s3_models_inference_service .namespace } '}}" ,
86+ metrics_query = f"pod:container_cpu_usage:sum{{namespace='{ ovms_kserve_inference_service .namespace } '}}" ,
9287 )
0 commit comments