@@ -5,6 +5,7 @@ set -exuo pipefail
55# Configuration
66WORK_DIR=" /tmp/llama-stack-integration-tests"
77INFERENCE_MODEL=" ${INFERENCE_MODEL:- Qwen/ Qwen3-0.6B} "
8+ EMBEDDING_MODEL=" ${EMBEDDING_MODEL:- nomic-embed-text-v1.5} "
89
910SCRIPT_DIR=" $( cd " $( dirname " ${BASH_SOURCE[0]} " ) " && pwd) "
1011
@@ -38,10 +39,15 @@ function clone_llama_stack() {
3839 cd " $WORK_DIR "
3940 # fetch origin incase we didn't clone a fresh repo
4041 git fetch origin
41- if ! git checkout " v$LLAMA_STACK_VERSION " ; then
42- echo " Error: Could not checkout tag v$LLAMA_STACK_VERSION "
42+ if [ " $LLAMA_STACK_VERSION " == " main" ]; then
43+ checkout_to=" main"
44+ else
45+ checkout_to=" v$LLAMA_STACK_VERSION "
46+ fi
47+ if ! git checkout " $checkout_to " ; then
48+ echo " Error: Could not checkout $checkout_to "
4349 echo " Available tags:"
44- git tag | grep " ^v " | tail -10
50+ git tag | tail -10
4551 exit 1
4652 fi
4753}
@@ -64,12 +70,13 @@ function run_integration_tests() {
6470 exit 1
6571 fi
6672
67- # TODO: remove this once we have a stable version of llama-stack client
68- # Currently, LLS client version is 0.3.0, while the server version is 0.3.0rc3+rhai0
69- uv run --with llama-stack-client==0.3.0 pytest -s -v tests/integration/inference/ \
73+ uv venv
74+ source .venv/bin/activate
75+ uv pip install llama-stack-client
76+ uv run pytest -s -v tests/integration/inference/ \
7077 --stack-config=server:" $STACK_CONFIG_PATH " \
7178 --text-model=vllm-inference/" $INFERENCE_MODEL " \
72- --embedding-model=granite-embedding-125m \
79+ --embedding-model=nomic-ai/ " $EMBEDDING_MODEL " \
7380 -k " not ($SKIP_TESTS )"
7481}
7582
0 commit comments