Skip to content

Commit 031e1b6

Browse files
committed
fix: served-model-name in vllm_cpu_runtime fixture
1 parent a49ee47 commit 031e1b6

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

tests/fixtures/inference.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
from ocp_resources.service import Service
1010
from ocp_resources.serving_runtime import ServingRuntime
1111

12-
from utilities.constants import RuntimeTemplates, KServeDeploymentType
12+
from utilities.constants import RuntimeTemplates, KServeDeploymentType, QWEN_MODEL_NAME
1313
from utilities.inference_utils import create_isvc
1414
from utilities.serving_runtime import ServingRuntimeFromTemplate
1515

@@ -35,6 +35,7 @@ def vllm_cpu_runtime(
3535
"args": [
3636
"--port=8032",
3737
"--model=/mnt/models",
38+
f"--served-model-name={QWEN_MODEL_NAME}",
3839
],
3940
"ports": [{"containerPort": 8032, "protocol": "TCP"}],
4041
"volumeMounts": [{"mountPath": "/dev/shm", "name": "shm"}],

0 commit comments

Comments
 (0)