Skip to content

Commit 5d06ccf

Browse files
authored
fix: guardrails ports, disable auth (for now) for LLS LMEval tests, fix eval api (#780)
1 parent b518461 commit 5d06ccf

File tree

4 files changed

+8
-8
lines changed

4 files changed

+8
-8
lines changed

tests/fixtures/inference.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,7 @@ def qwen_isvc(
6666
storage_key=minio_data_connection.name,
6767
storage_path="Qwen2.5-0.5B-Instruct",
6868
wait_for_predictor_pods=False,
69-
enable_auth=True,
69+
enable_auth=False,
7070
resources={
7171
"requests": {"cpu": "2", "memory": "10Gi"},
7272
"limits": {"cpu": "2", "memory": "12Gi"},
@@ -77,7 +77,7 @@ def qwen_isvc(
7777

7878
@pytest.fixture(scope="class")
7979
def qwen_isvc_url(qwen_isvc: InferenceService) -> str:
80-
return f"http://{qwen_isvc.name}-predictor.{qwen_isvc.namespace}.svc.cluster.local:8032/v1"
80+
return f"http://{qwen_isvc.name}-predictor.{qwen_isvc.namespace}.svc.cluster.local:80/v1"
8181

8282

8383
@pytest.fixture(scope="class")

tests/llama_stack/eval/test_lmeval_provider.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ def test_lmeval_register_benchmark(self, minio_pod, minio_data_connection, llama
5555
def test_llamastack_run_eval(
5656
self, minio_pod, minio_data_connection, patched_dsc_lmeval_allow_all, llama_stack_client
5757
):
58-
job = llama_stack_client.eval.run_eval(
58+
job = llama_stack_client.alpha.eval.run_eval(
5959
benchmark_id=TRUSTYAI_LMEVAL_ARCEASY,
6060
benchmark_config={
6161
"eval_candidate": {
@@ -65,14 +65,14 @@ def test_llamastack_run_eval(
6565
"sampling_params": {"temperature": 0.7, "top_p": 0.9, "max_tokens": 10},
6666
},
6767
"scoring_params": {},
68-
"num_examples": 2,
68+
"num_examples": 1,
6969
},
7070
)
7171

7272
samples = TimeoutSampler(
7373
wait_timeout=Timeout.TIMEOUT_10MIN,
7474
sleep=30,
75-
func=lambda: llama_stack_client.eval.jobs.status(
75+
func=lambda: llama_stack_client.alpha.eval.jobs.status(
7676
job_id=job.job_id, benchmark_id=TRUSTYAI_LMEVAL_ARCEASY
7777
).status,
7878
)

tests/model_explainability/guardrails/test_guardrails.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -244,7 +244,7 @@ def test_guardrails_builtin_detectors_negative_detection(
244244
"type": "text_contents",
245245
"service": {
246246
"hostname": f"{PROMPT_INJECTION_DETECTOR}-predictor",
247-
"port": 8000,
247+
"port": 80,
248248
},
249249
"chunker_id": "whole_doc_chunker",
250250
"default_threshold": 0.5,
@@ -253,7 +253,7 @@ def test_guardrails_builtin_detectors_negative_detection(
253253
"type": "text_contents",
254254
"service": {
255255
"hostname": f"{HAP_DETECTOR}-predictor",
256-
"port": 8000,
256+
"port": 80,
257257
},
258258
"chunker_id": "whole_doc_chunker",
259259
"default_threshold": 0.5,

utilities/constants.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -477,7 +477,7 @@ class LLMdInferenceSimConfig:
477477

478478

479479
LLM_D_CHAT_GENERATION_CONFIG: Dict[str, Any] = {
480-
"service": {"hostname": f"{LLMdInferenceSimConfig.isvc_name}-predictor", "port": 8032}
480+
"service": {"hostname": f"{LLMdInferenceSimConfig.isvc_name}-predictor", "port": 80}
481481
}
482482

483483

0 commit comments

Comments
 (0)