Skip to content

Commit f6dfab9

Browse files
committed
wip
1 parent b8798d4 commit f6dfab9

File tree

3 files changed

+14
-8
lines changed

3 files changed

+14
-8
lines changed

.github/workflows/redhat-distro-container.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ jobs:
7373
LLAMA_STACK_VERSION: ${{ env.LLAMA_STACK_COMMIT_SHA }}
7474
run: |
7575
tmp_build_dir=$(mktemp -d)
76-
git clone --filter=blob:none --no-checkout https://github.com/llamastack/llama-stack.git "$tmp_build_dir"
76+
git clone --filter=blob:none --no-checkout https://github.com/opendatahub-io/llama-stack.git "$tmp_build_dir"
7777
cd "$tmp_build_dir"
7878
git checkout "$LLAMA_STACK_VERSION"
7979
python3 -m venv .venv

tests/run_integration_tests.sh

Lines changed: 10 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -38,10 +38,15 @@ function clone_llama_stack() {
3838
cd "$WORK_DIR"
3939
# fetch origin incase we didn't clone a fresh repo
4040
git fetch origin
41-
if ! git checkout "v$LLAMA_STACK_VERSION"; then
42-
echo "Error: Could not checkout tag v$LLAMA_STACK_VERSION"
41+
if [ "$LLAMA_STACK_VERSION" == "main" ]; then
42+
checkout_to="main"
43+
else
44+
checkout_to="v$LLAMA_STACK_VERSION"
45+
fi
46+
if ! git checkout "$checkout_to"; then
47+
echo "Error: Could not checkout $checkout_to"
4348
echo "Available tags:"
44-
git tag | grep "^v" | tail -10
49+
git tag | tail -10
4550
exit 1
4651
fi
4752
}
@@ -64,9 +69,8 @@ function run_integration_tests() {
6469
exit 1
6570
fi
6671

67-
# TODO: remove this once we have a stable version of llama-stack client
68-
# Currently, LLS client version is 0.3.0, while the server version is 0.3.0rc3+rhai0
69-
uv run --with llama-stack-client==0.3.0 pytest -s -v tests/integration/inference/ \
72+
uv pip install llama-stack-client
73+
uv run pytest -s -v tests/integration/inference/ \
7074
--stack-config=server:"$STACK_CONFIG_PATH" \
7175
--text-model=vllm-inference/"$INFERENCE_MODEL" \
7276
--embedding-model=granite-embedding-125m \

tests/smoke.sh

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,7 @@ function test_model_list {
4242
return
4343
else
4444
echo "Model $INFERENCE_MODEL was not found :("
45+
echo "Response: $resp"
4546
echo "Container logs:"
4647
docker logs llama-stack || true
4748
exit 1
@@ -50,12 +51,13 @@ function test_model_list {
5051

5152
function test_model_openai_inference {
5253
echo "===> Attempting to chat with model $INFERENCE_MODEL..."
53-
resp=$(curl -fsS http://127.0.0.1:8321/v1/chat/completions -H "Content-Type: application/json" -d "{\"model\": \"$INFERENCE_MODEL\",\"messages\": [{\"role\": \"user\", \"content\": \"What color is grass?\"}], \"max_tokens\": 128, \"temperature\": 0.0}")
54+
resp=$(curl -fsS http://127.0.0.1:8321/v1/chat/completions -H "Content-Type: application/json" -d "{\"model\": \"vllm-inference/$INFERENCE_MODEL\",\"messages\": [{\"role\": \"user\", \"content\": \"What color is grass?\"}], \"max_tokens\": 128, \"temperature\": 0.0}")
5455
if echo "$resp" | grep -q "green"; then
5556
echo "===> Inference is working :)"
5657
return
5758
else
5859
echo "===> Inference is not working :("
60+
echo "Response: $resp"
5961
echo "Container logs:"
6062
docker logs llama-stack || true
6163
exit 1

0 commit comments

Comments
 (0)