File tree Expand file tree Collapse file tree 2 files changed +32
-1
lines changed
Expand file tree Collapse file tree 2 files changed +32
-1
lines changed Original file line number Diff line number Diff line change 1+ name : Setup VLLM
2+ description : Start VLLM
3+ runs :
4+ using : " composite"
5+ steps :
6+ - name : Start VLLM
7+ shell : bash
8+ run : |
9+ # Start vllm container
10+ docker run -d \
11+ --name vllm \
12+ -p 8000:8000 \
13+ --privileged=true \
14+ quay.io/higginsd/vllm-cpu:65393ee064 \
15+ --host 0.0.0.0 \
16+ --port 8000 \
17+ --enable-auto-tool-choice \
18+ --tool-call-parser llama3_json \
19+ --model /root/.cache/Llama-3.2-1B-Instruct \
20+ --served-model-name meta-llama/Llama-3.2-1B-Instruct
21+
22+ # Wait for vllm to be ready
23+ echo "Waiting for vllm to be ready..."
24+ timeout 900 bash -c 'until curl -f http://localhost:8000/health; do
25+ echo "Waiting for vllm..."
26+ sleep 5
27+ done'
Original file line number Diff line number Diff line change @@ -50,14 +50,18 @@ jobs:
5050 cache-from : type=gha
5151 cache-to : type=gha,mode=max
5252
53+ - name : Setup vllm for image test
54+ id : vllm
55+ uses : ./.github/actions/setup-vllm
56+
5357 - name : Test image
5458 id : test
5559 run : |
5660 set -euo pipefail
5761 # Start llama stack
5862 CID="$(docker run -d --pull=never \
5963 -p 8321:8321 \
60- --env INFERENCE_MODEL=dummy \
64+ --env INFERENCE_MODEL=meta-llama/Llama-3.2-1B-Instruct \
6165 --name llama-stack \
6266 ${{ env.IMAGE_NAME }}:${{ github.sha }})"
6367 trap 'docker rm -f "$CID" >/dev/null 2>&1 || true' EXIT
You can’t perform that action at this time.
0 commit comments