Skip to content

Commit 2730802

Browse files
committed
fix: use secretKeyRef for VLLM tokens to prevent credential leaks in logs
Move VLLM_API_TOKEN and VLLM_EMBEDDING_API_TOKEN from plain env var values to K8s secretKeyRef, matching the existing pattern used for POSTGRES_PASSWORD. Add both tokens to LLAMA_STACK_DISTRIBUTION_SECRET_DATA and relocate the dict after all secret variables are defined. Signed-off-by: Ignas Baranauskas <[email protected]>
1 parent cc43a76 commit 2730802

File tree

1 file changed

+21
-11
lines changed

1 file changed

+21
-11
lines changed

tests/llama_stack/conftest.py

Lines changed: 21 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -48,11 +48,6 @@
4848
POSTGRESQL_USER = os.getenv("LLS_VECTOR_IO_POSTGRESQL_USER", "ps_user")
4949
POSTGRESQL_PASSWORD = os.getenv("LLS_VECTOR_IO_POSTGRESQL_PASSWORD", "ps_password")
5050

51-
LLAMA_STACK_DISTRIBUTION_SECRET_DATA = {
52-
"postgres-user": POSTGRESQL_USER,
53-
"postgres-password": POSTGRESQL_PASSWORD,
54-
}
55-
5651
LLS_CORE_INFERENCE_MODEL = os.getenv("LLS_CORE_INFERENCE_MODEL", "")
5752
LLS_CORE_VLLM_URL = os.getenv("LLS_CORE_VLLM_URL", "")
5853
LLS_CORE_VLLM_API_TOKEN = os.getenv("LLS_CORE_VLLM_API_TOKEN", "")
@@ -68,6 +63,13 @@
6863
LLS_CORE_VLLM_EMBEDDING_MAX_TOKENS = os.getenv("LLS_CORE_VLLM_EMBEDDING_MAX_TOKENS", "8192")
6964
LLS_CORE_VLLM_EMBEDDING_TLS_VERIFY = os.getenv("LLS_CORE_VLLM_EMBEDDING_TLS_VERIFY", "true")
7065

66+
LLAMA_STACK_DISTRIBUTION_SECRET_DATA = {
67+
"postgres-user": POSTGRESQL_USER,
68+
"postgres-password": POSTGRESQL_PASSWORD,
69+
"vllm-api-token": LLS_CORE_VLLM_API_TOKEN,
70+
"vllm-embedding-api-token": LLS_CORE_VLLM_EMBEDDING_API_TOKEN,
71+
}
72+
7173
IBM_EARNINGS_DOC_URL = "https://www.ibm.com/downloads/documents/us-en/1550f7eea8c0ded6"
7274

7375
UPGRADE_DISTRIBUTION_NAME = "llama-stack-distribution-upgrade"
@@ -171,11 +173,12 @@ def test_with_remote_milvus(llama_stack_server_config):
171173
inference_model = LLS_CORE_INFERENCE_MODEL
172174
env_vars.append({"name": "INFERENCE_MODEL", "value": inference_model})
173175

174-
if params.get("vllm_api_token"):
175-
vllm_api_token = str(params.get("vllm_api_token"))
176-
else:
177-
vllm_api_token = LLS_CORE_VLLM_API_TOKEN
178-
env_vars.append({"name": "VLLM_API_TOKEN", "value": vllm_api_token})
176+
env_vars.append(
177+
{
178+
"name": "VLLM_API_TOKEN",
179+
"valueFrom": {"secretKeyRef": {"name": "llamastack-distribution-secret", "key": "vllm-api-token"}},
180+
},
181+
)
179182

180183
if params.get("vllm_url_fixture"):
181184
vllm_url = str(request.getfixturevalue(argname=params.get("vllm_url_fixture")))
@@ -200,7 +203,14 @@ def test_with_remote_milvus(llama_stack_server_config):
200203
env_vars.append({"name": "EMBEDDING_MODEL", "value": LLS_CORE_EMBEDDING_MODEL})
201204
env_vars.append({"name": "EMBEDDING_PROVIDER_MODEL_ID", "value": LLS_CORE_EMBEDDING_PROVIDER_MODEL_ID})
202205
env_vars.append({"name": "VLLM_EMBEDDING_URL", "value": LLS_CORE_VLLM_EMBEDDING_URL})
203-
env_vars.append({"name": "VLLM_EMBEDDING_API_TOKEN", "value": LLS_CORE_VLLM_EMBEDDING_API_TOKEN})
206+
env_vars.append(
207+
{
208+
"name": "VLLM_EMBEDDING_API_TOKEN",
209+
"valueFrom": {
210+
"secretKeyRef": {"name": "llamastack-distribution-secret", "key": "vllm-embedding-api-token"}
211+
},
212+
},
213+
)
204214
env_vars.append({"name": "VLLM_EMBEDDING_MAX_TOKENS", "value": LLS_CORE_VLLM_EMBEDDING_MAX_TOKENS})
205215
env_vars.append({"name": "VLLM_EMBEDDING_TLS_VERIFY", "value": LLS_CORE_VLLM_EMBEDDING_TLS_VERIFY})
206216
elif embedding_provider == "sentence-transformers":

0 commit comments

Comments
 (0)