Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@
from kubernetes.dynamic import DynamicClient
from ocp_resources.inference_service import InferenceService
from ocp_resources.namespace import Namespace
from ocp_resources.resource import ResourceEditor
from simple_logger.logger import get_logger

from tests.model_serving.model_runtime.vllm.basic_model_deployment.test_granite_7b_starter import SERVING_ARGUMENT
Expand All @@ -17,7 +16,6 @@
)
from utilities.constants import ModelFormat, ModelVersion, Protocols, RunTimeConfigs, Timeout
from utilities.inference_utils import Inference
from utilities.jira import is_jira_open
from utilities.manifests.onnx import ONNX_INFERENCE_CONFIG
from utilities.monitoring import validate_metrics_field

Expand Down Expand Up @@ -74,31 +72,6 @@ def test_ovms_keda_scaling_verify_scaledobject(
admin_client: DynamicClient,
):
"""Test KEDA ScaledObject configuration and run inference multiple times to trigger scaling."""

if is_jira_open(jira_id="RHOAIENG-31386", admin_client=admin_client):
isvc_dict = stressed_ovms_keda_inference_service.instance.to_dict()
metrics = isvc_dict.get("spec", {}).get("predictor", {}).get("autoScaling", {}).get("metrics", [])

if metrics and isinstance(metrics[0], dict) and metrics[0].get("external") is not None:
metrics[0].setdefault("external", {})["authenticationRef"] = {
"authModes": "bearer",
"authenticationRef": {"name": "inference-prometheus-auth"},
}

ResourceEditor(
patches={
stressed_ovms_keda_inference_service: {
"spec": {
"predictor": {
"autoScaling": {
"metrics": metrics,
}
}
}
}
}
).update()

verify_keda_scaledobject(
client=unprivileged_client,
isvc=stressed_ovms_keda_inference_service,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,6 @@ def test_rest_raw_deployment_internal_route(self, s3_models_inference_service):
use_default_query=True,
)

@pytest.mark.jira("RHOAIENG-17322", run=False)
@pytest.mark.parametrize(
"patched_s3_caikit_kserve_isvc_visibility_label",
[
Expand Down Expand Up @@ -190,8 +189,7 @@ def test_rest_raw_deployment_exposed_route_with_timeout(self, s3_models_inferenc
],
indirect=True,
)
@pytest.mark.skip(reason="skipping grpc raw for tgis-caikit RHOAIENG-17783")
Comment thread
dbasunag marked this conversation as resolved.
@pytest.mark.jira("RHOAIENG-17783", run=False)
@pytest.mark.skip(reason="skipping grpc raw for tgis-caikit")
Comment thread
dbasunag marked this conversation as resolved.
class TestGrpcRawDeployment:
def test_grpc_raw_deployment_internal_route(self, s3_models_inference_service):
"""Test GRPC inference using internal route"""
Expand Down Expand Up @@ -247,8 +245,7 @@ def test_grpc_raw_deployment_exposed_route(self, patched_s3_caikit_kserve_isvc_v
],
indirect=True,
)
@pytest.mark.skip(reason="skipping grpc raw for tgis-caikit RHOAIENG-17783")
@pytest.mark.jira("RHOAIENG-17783", run=False)
@pytest.mark.skip(reason="skipping grpc raw for tgis-caikit")
class TestGrpcRawDeploymentTimeout:
@pytest.mark.dependency(name="test_grpc_raw_deployment_exposed_route")
def test_grpc_raw_deployment_exposed_route(self, s3_models_inference_service):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ def multi_node_serving_runtime(
) -> Generator[ServingRuntime, Any, Any]:
with ServingRuntimeFromTemplate(
client=unprivileged_client,
name="vllm-multinode-runtime", # TODO: rename servingruntime when RHOAIENG-16147 is resolved
name="vllm-multinode-runtime",
namespace=unprivileged_model_namespace.name,
template_name="vllm-multinode-runtime-template",
multi_model=False,
Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,5 @@
"""Tests for invalid model name in inference endpoint.

Jira: RHOAIENG-48282
"""
Tests for invalid model name in inference endpoint.
"""

import json
Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,5 @@
"""Tests for malformed JSON payload handling in inference requests.

Jira: RHOAIENG-48279
"""
Tests for malformed JSON payload handling in inference requests.
"""

from http import HTTPStatus
Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,5 @@
"""Tests for missing required fields in inference requests.

Jira: RHOAIENG-48281
"""
Tests for missing required fields in inference requests.
"""

import json
Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,5 @@
"""Tests for unsupported Content-Type headers in inference requests.

Jira: RHOAIENG-48283
"""
Tests for unsupported Content-Type headers in inference requests.
"""

import json
Expand Down
Original file line number Diff line number Diff line change
@@ -1,6 +1,5 @@
"""Tests for wrong data types in input tensor.

Jira: RHOAIENG-48280
"""
Tests for wrong data types in input tensor.
"""

import copy
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,6 @@ def wait_for_isvc_model_status(isvc: InferenceService, target_model_state: str,
raise


@pytest.mark.jira("RHOAIENG-10765")
@pytest.mark.parametrize(
"unprivileged_model_namespace, serving_runtime_from_template, invalid_s3_models_inference_service",
[
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,6 @@
)
class TestKserveModelCar:
@pytest.mark.tier1
@pytest.mark.jira("RHOAIENG-13465")
def test_model_car_no_restarts(self, model_car_inference_service):
"""Verify that model pod doesn't restart"""
pod = get_pods_by_isvc_label(
Expand All @@ -56,7 +55,7 @@ def test_model_car_no_restarts(self, model_car_inference_service):

@pytest.mark.tier1
@pytest.mark.ocp_interop
@pytest.mark.jira("RHOAIENG-12306")
Comment thread
dbasunag marked this conversation as resolved.
@pytest.mark.skip(reason="Skipping due to RHOAIENG-12306")
def test_model_car_using_rest(self, model_car_inference_service):
"""Verify model query with token using REST"""
verify_inference_response(
Expand All @@ -69,7 +68,7 @@ def test_model_car_using_rest(self, model_car_inference_service):

@pytest.mark.tier1
@pytest.mark.ocp_interop
@pytest.mark.jira("RHOAIENG-38674")
Comment thread
dbasunag marked this conversation as resolved.
@pytest.mark.skip(reason="Skipping due to RHOAIENG-38674")
def test_model_status_loaded(self, model_car_inference_service):
"""Verify model status on the InferenceService resource is in a valid state."""
model_status = model_car_inference_service.instance.status.modelStatus
Expand Down
Loading