77 - " .github/workflows/pr-sglang-g6-inference.yaml"
88
99env :
10- SGLANG_IMAGE : " lmsysorg/ sglang:v0 .5.7-runtime "
10+ SGLANG_IMAGE : " public.ecr.aws/deep-learning-containers/ sglang:0 .5.5-gpu-py312 "
1111 SGLANG_VERSION : " 0.5.7"
12- DATASET_URL : " https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json"
1312
1413jobs :
1514 sglang-upstream-test :
@@ -21,38 +20,46 @@ jobs:
2120 - name : Pull image
2221 run : docker pull ${{ env.SGLANG_IMAGE }}
2322
24- - name : Checkout SGLang tests
25- uses : actions/checkout@v5
26- with :
27- repository : sgl-project/sglang
28- ref : v${{ env.SGLANG_VERSION }}
29- path : sglang_source
30-
23+ - name : Setup for SGLang datasets
24+ run : |
25+ mkdir -p ${HOME}/dataset
26+ if [ ! -f ${HOME}/dataset/ShareGPT_V3_unfiltered_cleaned_split.json ]; then
27+ echo "Downloading ShareGPT dataset..."
28+ wget -P ${HOME}/dataset https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json
29+ else
30+ echo "ShareGPT dataset already exists. Skipping download."
31+ fi
32+
3133 - name : Start container
3234 run : |
33- CONTAINER_ID=$(docker run -d --gpus=all --entrypoint /bin/bash \
35+ CONTAINER_ID=$(docker run -d -it --rm --gpus=all \
3436 -v ${HOME}/.cache/huggingface:/root/.cache/huggingface \
35- -v ./sglang_source:/workdir --workdir /workdir \
37+ -v ${HOME}/dataset:/dataset \
38+ -p 30000:30000 \
39+ -e SM_SGLANG_MODEL_PATH=Qwen/Qwen3-0.6B \
40+ -e SM_SGLANG_REASONING_PARSER=qwen3 \
41+ -e SM_SGLANG_HOST=127.0.0.1 \
42+ -e SM_SGLANG_PORT=30000 \
3643 -e HF_TOKEN=${{ secrets.HUGGING_FACE_HUB_TOKEN }} \
37- ${{ env.SGLANG_IMAGE }})
44+ ${{ env.SGLANG_IMAGE }}
3845 echo "CONTAINER_ID=${CONTAINER_ID}" >> ${GITHUB_ENV}
39-
40- - name : Setup for SGLang tests
41- run : |
42- docker exec ${CONTAINER_ID} sh -c 'set -eux
43- bash scripts/ci/ci_install_dependency.sh'
44-
45- - name : Run SGLang upstream tests
46+ echo "Waiting for serving endpoint startup ..."
47+ sleep 120s
48+ docker logs ${CONTAINER_ID}
49+
50+ - name : Run SGLang tests
4651 run : |
47- docker exec ${CONTAINER_ID} sh -c 'set -eux
48- nvidia-smi
49- # SRT backend Test
50- cd /workdir/test
51- python3 run_suite.py --hw cuda --suite stage-a-test-1'
52+ docker exec ${CONTAINER_ID} python3 -m sglang.bench_serving \
53+ --backend sglang \
54+ --host 127.0.0.1 --port 30000 \
55+ --num-prompts 1000 \
56+ --model Qwen/Qwen3-0.6B \
57+ --dataset-name sharegpt \
58+ --dataset-path /dataset/ShareGPT_V3_unfiltered_cleaned_split.json
5259
5360 - name : Show container logs
5461 if : always()
55- run : docker logs ${CONTAINER_ID} || true
62+ run : docker logs ${CONTAINER_ID}
5663
5764 - name : Cleanup
5865 if : always()
0 commit comments