Skip to content

Commit 6af7265

Browse files
committed
fix: various fixes
Signed-off-by: Sébastien Han <seb@redhat.com>
1 parent b8798d4 commit 6af7265

File tree

3 files changed

+19
-10
lines changed

3 files changed

+19
-10
lines changed

.github/workflows/redhat-distro-container.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ jobs:
4444
runs-on: ubuntu-latest
4545
env:
4646
INFERENCE_MODEL: Qwen/Qwen3-0.6B
47-
EMBEDDING_MODEL: granite-embedding-125m
47+
EMBEDDING_MODEL: nomic-embed-text-v1.5
4848
VLLM_URL: http://localhost:8000/v1
4949
LLAMA_STACK_COMMIT_SHA: ${{ github.event.inputs.llama_stack_commit_sha || 'main' }}
5050
strategy:
@@ -73,7 +73,7 @@ jobs:
7373
LLAMA_STACK_VERSION: ${{ env.LLAMA_STACK_COMMIT_SHA }}
7474
run: |
7575
tmp_build_dir=$(mktemp -d)
76-
git clone --filter=blob:none --no-checkout https://github.com/llamastack/llama-stack.git "$tmp_build_dir"
76+
git clone --filter=blob:none --no-checkout https://github.com/opendatahub-io/llama-stack.git "$tmp_build_dir"
7777
cd "$tmp_build_dir"
7878
git checkout "$LLAMA_STACK_VERSION"
7979
python3 -m venv .venv

tests/run_integration_tests.sh

Lines changed: 14 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,7 @@ set -exuo pipefail
55
# Configuration
66
WORK_DIR="/tmp/llama-stack-integration-tests"
77
INFERENCE_MODEL="${INFERENCE_MODEL:-Qwen/Qwen3-0.6B}"
8+
EMBEDDING_MODEL="${EMBEDDING_MODEL:-nomic-embed-text-v1.5}"
89

910
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
1011

@@ -38,10 +39,15 @@ function clone_llama_stack() {
3839
cd "$WORK_DIR"
3940
# fetch origin incase we didn't clone a fresh repo
4041
git fetch origin
41-
if ! git checkout "v$LLAMA_STACK_VERSION"; then
42-
echo "Error: Could not checkout tag v$LLAMA_STACK_VERSION"
42+
if [ "$LLAMA_STACK_VERSION" == "main" ]; then
43+
checkout_to="main"
44+
else
45+
checkout_to="v$LLAMA_STACK_VERSION"
46+
fi
47+
if ! git checkout "$checkout_to"; then
48+
echo "Error: Could not checkout $checkout_to"
4349
echo "Available tags:"
44-
git tag | grep "^v" | tail -10
50+
git tag | tail -10
4551
exit 1
4652
fi
4753
}
@@ -64,12 +70,13 @@ function run_integration_tests() {
6470
exit 1
6571
fi
6672

67-
# TODO: remove this once we have a stable version of llama-stack client
68-
# Currently, LLS client version is 0.3.0, while the server version is 0.3.0rc3+rhai0
69-
uv run --with llama-stack-client==0.3.0 pytest -s -v tests/integration/inference/ \
73+
uv venv
74+
source .venv/bin/activate
75+
uv pip install llama-stack-client
76+
uv run pytest -s -v tests/integration/inference/ \
7077
--stack-config=server:"$STACK_CONFIG_PATH" \
7178
--text-model=vllm-inference/"$INFERENCE_MODEL" \
72-
--embedding-model=granite-embedding-125m \
79+
--embedding-model=nomic-ai/"$EMBEDDING_MODEL" \
7380
-k "not ($SKIP_TESTS)"
7481
}
7582

tests/smoke.sh

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,7 @@ function test_model_list {
4242
return
4343
else
4444
echo "Model $INFERENCE_MODEL was not found :("
45+
echo "Response: $resp"
4546
echo "Container logs:"
4647
docker logs llama-stack || true
4748
exit 1
@@ -50,12 +51,13 @@ function test_model_list {
5051

5152
function test_model_openai_inference {
5253
echo "===> Attempting to chat with model $INFERENCE_MODEL..."
53-
resp=$(curl -fsS http://127.0.0.1:8321/v1/chat/completions -H "Content-Type: application/json" -d "{\"model\": \"$INFERENCE_MODEL\",\"messages\": [{\"role\": \"user\", \"content\": \"What color is grass?\"}], \"max_tokens\": 128, \"temperature\": 0.0}")
54+
resp=$(curl -fsS http://127.0.0.1:8321/v1/chat/completions -H "Content-Type: application/json" -d "{\"model\": \"vllm-inference/$INFERENCE_MODEL\",\"messages\": [{\"role\": \"user\", \"content\": \"What color is grass?\"}], \"max_tokens\": 128, \"temperature\": 0.0}")
5455
if echo "$resp" | grep -q "green"; then
5556
echo "===> Inference is working :)"
5657
return
5758
else
5859
echo "===> Inference is not working :("
60+
echo "Response: $resp"
5961
echo "Container logs:"
6062
docker logs llama-stack || true
6163
exit 1

0 commit comments

Comments
 (0)