Skip to content

Commit 9311f60

Browse files
migrate to ovms (#904)
* remove kueue test from smoek 2.22 Signed-off-by: Milind Waykole <mwaykole@redhat.com> * Fix tierdown deletion of project and migrate to ovms test from caikit Signed-off-by: Milind Waykole <mwaykole@redhat.com> * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fix: remove unused Caikit fixtures after OVMS migration * feat: convert remaining Caikit smoke tests to OVMS MNIST - Convert test_model_metrics.py from CAIKIT_TGIS to OVMS ONNX - Convert test_bge_large_eng_caikit.py to test_ovms_mnist_raw.py - Update metrics queries from tgi_* to ovms_* - Use ONNX_INFERENCE_CONFIG instead of CAIKIT_* configs --------- Signed-off-by: Milind Waykole <mwaykole@redhat.com> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
1 parent 6cd0d1f commit 9311f60

File tree

3 files changed

+63
-87
lines changed

3 files changed

+63
-87
lines changed

tests/model_serving/model_server/metrics/test_model_metrics.py

Lines changed: 14 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -7,13 +7,12 @@
77
from utilities.constants import (
88
KServeDeploymentType,
99
ModelFormat,
10-
ModelInferenceRuntime,
11-
ModelStoragePath,
10+
ModelVersion,
1211
Protocols,
1312
RuntimeTemplates,
1413
)
1514
from utilities.inference_utils import Inference
16-
from utilities.manifests.caikit_tgis import CAIKIT_TGIS_INFERENCE_CONFIG
15+
from utilities.manifests.onnx import ONNX_INFERENCE_CONFIG
1716
from utilities.monitoring import get_metrics_value, validate_metrics_field
1817

1918
pytestmark = [
@@ -28,17 +27,18 @@
2827
"unprivileged_model_namespace, serving_runtime_from_template, s3_models_inference_service",
2928
[
3029
pytest.param(
31-
{"name": "kserve-tgis-metrics"},
30+
{"name": "kserve-ovms-metrics"},
3231
{
33-
"name": f"{Protocols.HTTP}-{ModelInferenceRuntime.CAIKIT_TGIS_RUNTIME}",
34-
"template-name": RuntimeTemplates.CAIKIT_TGIS_SERVING,
32+
"name": f"{Protocols.HTTP}-{ModelFormat.ONNX}",
33+
"template-name": RuntimeTemplates.OVMS_KSERVE,
3534
"multi-model": False,
3635
"enable-http": True,
3736
},
3837
{
39-
"name": f"{Protocols.HTTP}-{ModelFormat.CAIKIT}",
38+
"name": f"{Protocols.HTTP}-{ModelFormat.ONNX}",
4039
"deployment-mode": KServeDeploymentType.SERVERLESS,
41-
"model-dir": ModelStoragePath.FLAN_T5_SMALL_CAIKIT,
40+
"model-dir": "test-dir",
41+
"model-version": ModelVersion.OPSET13,
4242
},
4343
)
4444
],
@@ -51,15 +51,14 @@ def test_model_metrics_num_success_requests(self, s3_models_inference_service, p
5151
"""Verify number of successful model requests in OpenShift monitoring system (UserWorkloadMonitoring) metrics"""
5252
verify_inference_response(
5353
inference_service=s3_models_inference_service,
54-
inference_config=CAIKIT_TGIS_INFERENCE_CONFIG,
55-
inference_type=Inference.ALL_TOKENS,
54+
inference_config=ONNX_INFERENCE_CONFIG,
55+
inference_type=Inference.INFER,
5656
protocol=Protocols.HTTPS,
57-
model_name=ModelFormat.CAIKIT,
5857
use_default_query=True,
5958
)
6059
validate_metrics_field(
6160
prometheus=prometheus,
62-
metrics_query="tgi_request_success",
61+
metrics_query="ovms_requests_success",
6362
expected_value="1",
6463
)
6564

@@ -71,16 +70,15 @@ def test_model_metrics_num_total_requests(self, s3_models_inference_service, pro
7170

7271
run_inference_multiple_times(
7372
isvc=s3_models_inference_service,
74-
inference_config=CAIKIT_TGIS_INFERENCE_CONFIG,
75-
inference_type=Inference.ALL_TOKENS,
73+
inference_config=ONNX_INFERENCE_CONFIG,
74+
inference_type=Inference.INFER,
7675
protocol=Protocols.HTTPS,
77-
model_name=ModelFormat.CAIKIT,
7876
iterations=total_runs,
7977
run_in_parallel=True,
8078
)
8179
validate_metrics_field(
8280
prometheus=prometheus,
83-
metrics_query="tgi_request_count",
81+
metrics_query="ovms_requests_success",
8482
expected_value=str(total_runs + 1),
8583
)
8684

tests/model_serving/model_server/raw_deployment/test_bge_large_eng_caikit.py

Lines changed: 0 additions & 71 deletions
This file was deleted.
Lines changed: 49 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,49 @@
1+
import pytest
2+
3+
from tests.model_serving.model_server.utils import verify_inference_response
4+
from utilities.constants import (
5+
KServeDeploymentType,
6+
ModelFormat,
7+
ModelVersion,
8+
Protocols,
9+
RuntimeTemplates,
10+
)
11+
from utilities.inference_utils import Inference
12+
from utilities.manifests.onnx import ONNX_INFERENCE_CONFIG
13+
14+
pytestmark = [pytest.mark.rawdeployment, pytest.mark.usefixtures("valid_aws_config")]
15+
16+
17+
@pytest.mark.smoke
18+
@pytest.mark.jira("RHOAIENG-11749")
19+
@pytest.mark.parametrize(
20+
"unprivileged_model_namespace, serving_runtime_from_template, s3_models_inference_service",
21+
[
22+
pytest.param(
23+
{"name": "raw-deployment-ovms-mnist"},
24+
{
25+
"name": f"{Protocols.HTTP}-{ModelFormat.ONNX}",
26+
"template-name": RuntimeTemplates.OVMS_KSERVE,
27+
"multi-model": False,
28+
"enable-http": True,
29+
},
30+
{
31+
"name": f"{Protocols.HTTP}-{ModelFormat.ONNX}",
32+
"deployment-mode": KServeDeploymentType.RAW_DEPLOYMENT,
33+
"model-dir": "test-dir",
34+
"model-version": ModelVersion.OPSET13,
35+
},
36+
)
37+
],
38+
indirect=True,
39+
)
40+
class TestOvmsMnistRaw:
41+
def test_ovms_mnist_inference_raw_internal_route(self, s3_models_inference_service):
42+
"""Test OVMS MNIST model inference using internal route"""
43+
verify_inference_response(
44+
inference_service=s3_models_inference_service,
45+
inference_config=ONNX_INFERENCE_CONFIG,
46+
inference_type=Inference.INFER,
47+
protocol=Protocols.HTTP,
48+
use_default_query=True,
49+
)

0 commit comments

Comments
 (0)