@@ -113,190 +113,194 @@ jobs:
113113 tag-pr : ${{ steps.config.outputs.framework }}-${{ steps.config.outputs.framework-version }}-${{ steps.config.outputs.device-type }}-${{ steps.config.outputs.python-version }}-${{ steps.config.outputs.cuda-version }}-${{ steps.config.outputs.os-version }}-rayserve-ec2-pr-${{ github.event.pull_request.number }}
114114 dockerfile-path : docker/${{ steps.config.outputs.framework }}/Dockerfile
115115
116- set-rayserve-test-environment :
117- needs : [build-vllm-rayserve-image, load-config]
118- if : |
119- always() && !failure() && !cancelled()
120- runs-on : ubuntu-latest
121- concurrency :
122- group : ${{ github.workflow }}-set-rayserve-test-environment-${{ github.event.pull_request.number }}
123- cancel-in-progress : true
124- outputs :
125- aws-account-id : ${{ steps.set-env.outputs.AWS_ACCOUNT_ID }}
126- image-uri : ${{ steps.set-env.outputs.IMAGE_URI }}
127- framework-version : ${{ steps.config.outputs.framework-version }}
128- steps :
129- - name : Checkout code
130- uses : actions/checkout@v4
131-
132- - name : Parse RayServe config
133- id : config
134- run : |
135- echo '${{ needs.load-config.outputs.rayserve-config }}' > config.json
136- echo "framework-version=$(jq -r '.common.framework_version' config.json)" >> $GITHUB_OUTPUT
137- echo "prod-image=$(jq -r '.common.prod_image' config.json)" >> $GITHUB_OUTPUT
138-
139- - name : Set test environment
140- id : set-env
141- run : |
142- if [[ "${{ needs.build-vllm-rayserve-image.result }}" == "success" ]]; then
143- AWS_ACCOUNT_ID=${{ vars.CI_AWS_ACCOUNT_ID }}
144- IMAGE_URI=${{ needs.build-vllm-rayserve-image.outputs.ci-image }}
145- else
146- AWS_ACCOUNT_ID=${{ vars.PROD_AWS_ACCOUNT_ID }}
147- IMAGE_URI=${{ vars.PROD_AWS_ACCOUNT_ID }}.dkr.ecr.${{ vars.AWS_REGION }}.amazonaws.com/${{ steps.config.outputs.prod-image }}
148- fi
149-
150- echo "Image URI to test: ${IMAGE_URI}"
151- echo "AWS_ACCOUNT_ID=${AWS_ACCOUNT_ID}" >> ${GITHUB_OUTPUT}
152- echo "IMAGE_URI=${IMAGE_URI}" >> ${GITHUB_OUTPUT}
153-
154- vllm-rayserve-regression-test :
155- needs : [build-vllm-rayserve-image, set-rayserve-test-environment]
156- if : success()
157- runs-on :
158- - codebuild-runner-${{ github.run_id }}-${{ github.run_attempt }}
159- fleet:x86-g6xl-runner
160- buildspec-override:true
161- concurrency :
162- group : ${{ github.workflow }}-vllm-rayserve-regression-test-${{ github.event.pull_request.number }}
163- cancel-in-progress : true
164- steps :
165- - name : Checkout DLC source
166- uses : actions/checkout@v5
167-
168- - name : Container pull
169- uses : ./.github/actions/ecr-authenticate
170- with :
171- aws-account-id : ${{ needs.set-rayserve-test-environment.outputs.aws-account-id }}
172- aws-region : ${{ vars.AWS_REGION }}
173- image-uri : ${{ needs.set-rayserve-test-environment.outputs.image-uri }}
174-
175- - name : Checkout vLLM tests
176- uses : actions/checkout@v5
177- with :
178- repository : vllm-project/vllm
179- ref : v${{ needs.set-rayserve-test-environment.outputs.framework-version }}
180- path : vllm_source
181-
182- - name : Start container
183- run : |
184- CONTAINER_ID=$(docker run -d -it --rm --gpus=all --entrypoint /bin/bash \
185- -v ${HOME}/.cache/huggingface:/root/.cache/huggingface \
186- -v ${HOME}/.cache/vllm:/root/.cache/vllm \
187- -v .:/workdir --workdir /workdir \
188- -e HF_TOKEN=${{ secrets.HUGGING_FACE_HUB_TOKEN }} \
189- ${{ needs.set-rayserve-test-environment.outputs.image-uri }})
190- echo "CONTAINER_ID=$CONTAINER_ID" >> $GITHUB_ENV
191-
192- - name : Setup for vLLM tests
193- run : |
194- docker exec ${CONTAINER_ID} scripts/vllm/vllm_0_10_2_test_setup.sh
195-
196- - name : Run vLLM tests
197- run : |
198- docker exec ${CONTAINER_ID} scripts/vllm/vllm_regression_test.sh
199-
200- vllm-rayserve-cuda-test :
201- needs : [build-vllm-rayserve-image, set-rayserve-test-environment]
202- if : success()
203- runs-on :
204- - codebuild-runner-${{ github.run_id }}-${{ github.run_attempt }}
205- fleet:x86-g6xl-runner
206- buildspec-override:true
207- concurrency :
208- group : ${{ github.workflow }}-vllm-rayserve-cuda-test-${{ github.event.pull_request.number }}
209- cancel-in-progress : true
210- steps :
211- - name : Checkout DLC source
212- uses : actions/checkout@v5
213-
214- - name : Container pull
215- uses : ./.github/actions/ecr-authenticate
216- with :
217- aws-account-id : ${{ needs.set-rayserve-test-environment.outputs.aws-account-id }}
218- aws-region : ${{ vars.AWS_REGION }}
219- image-uri : ${{ needs.set-rayserve-test-environment.outputs.image-uri }}
220-
221- - name : Checkout vLLM tests
222- uses : actions/checkout@v5
223- with :
224- repository : vllm-project/vllm
225- ref : v${{ needs.set-rayserve-test-environment.outputs.framework-version }}
226- path : vllm_source
227-
228- - name : Start container
229- run : |
230- CONTAINER_ID=$(docker run -d -it --rm --gpus=all --entrypoint /bin/bash \
231- -v ${HOME}/.cache/huggingface:/root/.cache/huggingface \
232- -v ${HOME}/.cache/vllm:/root/.cache/vllm \
233- -v .:/workdir --workdir /workdir \
234- -e HF_TOKEN=${{ secrets.HUGGING_FACE_HUB_TOKEN }} \
235- ${{ needs.set-rayserve-test-environment.outputs.image-uri }})
236- echo "CONTAINER_ID=$CONTAINER_ID" >> $GITHUB_ENV
237-
238- - name : Setup for vLLM tests
239- run : |
240- docker exec ${CONTAINER_ID} scripts/vllm/vllm_0_10_2_test_setup.sh
241-
242- - name : Run vLLM tests
243- run : |
244- docker exec ${CONTAINER_ID} scripts/vllm/vllm_cuda_test.sh
245-
246- vllm-rayserve-example-test :
247- needs : [build-vllm-rayserve-image, set-rayserve-test-environment]
248- if : success()
249- runs-on :
250- - codebuild-runner-${{ github.run_id }}-${{ github.run_attempt }}
251- fleet:x86-g6xl-runner
252- buildspec-override:true
253- concurrency :
254- group : ${{ github.workflow }}-vllm-rayserve-example-test-${{ github.event.pull_request.number }}
255- cancel-in-progress : true
256- steps :
257- - name : Checkout DLC source
258- uses : actions/checkout@v5
259-
260- - name : Container pull
261- uses : ./.github/actions/ecr-authenticate
262- with :
263- aws-account-id : ${{ needs.set-rayserve-test-environment.outputs.aws-account-id }}
264- aws-region : ${{ vars.AWS_REGION }}
265- image-uri : ${{ needs.set-rayserve-test-environment.outputs.image-uri }}
266-
267- - name : Checkout vLLM tests
268- uses : actions/checkout@v5
269- with :
270- repository : vllm-project/vllm
271- ref : v${{ needs.set-rayserve-test-environment.outputs.framework-version }}
272- path : vllm_source
273-
274- - name : Start container
275- run : |
276- CONTAINER_ID=$(docker run -d -it --rm --gpus=all --entrypoint /bin/bash \
277- -v ${HOME}/.cache/huggingface:/root/.cache/huggingface \
278- -v ${HOME}/.cache/vllm:/root/.cache/vllm \
279- -v .:/workdir --workdir /workdir \
280- -e HF_TOKEN=${{ secrets.HUGGING_FACE_HUB_TOKEN }} \
281- ${{ needs.set-rayserve-test-environment.outputs.image-uri }})
282- echo "CONTAINER_ID=$CONTAINER_ID" >> $GITHUB_ENV
283-
284- - name : Setup for vLLM tests
285- run : |
286- docker exec ${CONTAINER_ID} scripts/vllm/vllm_0_10_2_test_setup.sh
287-
288- - name : Run vLLM tests
289- run : |
290- docker exec ${CONTAINER_ID} scripts/vllm/vllm_rayserve_examples_test.sh
116+ # set-rayserve-test-environment:
117+ # needs: [build-vllm-rayserve-image, load-config]
118+ # if: |
119+ # always() && !failure() && !cancelled()
120+ # runs-on: ubuntu-latest
121+ # concurrency:
122+ # group: ${{ github.workflow }}-set-rayserve-test-environment-${{ github.event.pull_request.number }}
123+ # cancel-in-progress: true
124+ # outputs:
125+ # aws-account-id: ${{ steps.set-env.outputs.AWS_ACCOUNT_ID }}
126+ # image-uri: ${{ steps.set-env.outputs.IMAGE_URI }}
127+ # framework-version: ${{ steps.config.outputs.framework-version }}
128+ # steps:
129+ # - name: Checkout code
130+ # uses: actions/checkout@v4
131+
132+ # - name: Parse RayServe config
133+ # id: config
134+ # run: |
135+ # echo '${{ needs.load-config.outputs.rayserve-config }}' > config.json
136+ # echo "framework-version=$(jq -r '.common.framework_version' config.json)" >> $GITHUB_OUTPUT
137+ # echo "prod-image=$(jq -r '.common.prod_image' config.json)" >> $GITHUB_OUTPUT
138+
139+ # - name: Set test environment
140+ # id: set-env
141+ # run: |
142+ # if [[ "${{ needs.build-vllm-rayserve-image.result }}" == "success" ]]; then
143+ # AWS_ACCOUNT_ID=${{ vars.CI_AWS_ACCOUNT_ID }}
144+ # IMAGE_URI=${{ needs.build-vllm-rayserve-image.outputs.ci-image }}
145+ # else
146+ # AWS_ACCOUNT_ID=${{ vars.PROD_AWS_ACCOUNT_ID }}
147+ # IMAGE_URI=${{ vars.PROD_AWS_ACCOUNT_ID }}.dkr.ecr.${{ vars.AWS_REGION }}.amazonaws.com/${{ steps.config.outputs.prod-image }}
148+ # fi
149+
150+ # echo "Image URI to test: ${IMAGE_URI}"
151+ # echo "AWS_ACCOUNT_ID=${AWS_ACCOUNT_ID}" >> ${GITHUB_OUTPUT}
152+ # echo "IMAGE_URI=${IMAGE_URI}" >> ${GITHUB_OUTPUT}
153+
154+ # vllm-rayserve-regression-test:
155+ # needs: [build-vllm-rayserve-image, set-rayserve-test-environment]
156+ # if: success()
157+ # runs-on:
158+ # - codebuild-runner-${{ github.run_id }}-${{ github.run_attempt }}
159+ # fleet:x86-g6xl-runner
160+ # buildspec-override:true
161+ # concurrency:
162+ # group: ${{ github.workflow }}-vllm-rayserve-regression-test-${{ github.event.pull_request.number }}
163+ # cancel-in-progress: true
164+ # steps:
165+ # - name: Checkout DLC source
166+ # uses: actions/checkout@v5
167+
168+ # - name: Container pull
169+ # uses: ./.github/actions/ecr-authenticate
170+ # with:
171+ # aws-account-id: ${{ needs.set-rayserve-test-environment.outputs.aws-account-id }}
172+ # aws-region: ${{ vars.AWS_REGION }}
173+ # image-uri: ${{ needs.set-rayserve-test-environment.outputs.image-uri }}
174+
175+ # - name: Checkout vLLM tests
176+ # uses: actions/checkout@v5
177+ # with:
178+ # repository: vllm-project/vllm
179+ # ref: v${{ needs.set-rayserve-test-environment.outputs.framework-version }}
180+ # path: vllm_source
181+
182+ # - name: Start container
183+ # run: |
184+ # CONTAINER_ID=$(docker run -d -it --rm --gpus=all --entrypoint /bin/bash \
185+ # -v ${HOME}/.cache/huggingface:/root/.cache/huggingface \
186+ # -v ${HOME}/.cache/vllm:/root/.cache/vllm \
187+ # -v .:/workdir --workdir /workdir \
188+ # -e HF_TOKEN=${{ secrets.HUGGING_FACE_HUB_TOKEN }} \
189+ # ${{ needs.set-rayserve-test-environment.outputs.image-uri }})
190+ # echo "CONTAINER_ID=$CONTAINER_ID" >> $GITHUB_ENV
191+
192+ # - name: Setup for vLLM tests
193+ # run: |
194+ # docker exec ${CONTAINER_ID} scripts/vllm/vllm_0_10_2_test_setup.sh
195+
196+ # - name: Run vLLM tests
197+ # run: |
198+ # docker exec ${CONTAINER_ID} scripts/vllm/vllm_regression_test.sh
199+
200+ # vllm-rayserve-cuda-test:
201+ # needs: [build-vllm-rayserve-image, set-rayserve-test-environment]
202+ # if: success()
203+ # runs-on:
204+ # - codebuild-runner-${{ github.run_id }}-${{ github.run_attempt }}
205+ # fleet:x86-g6xl-runner
206+ # buildspec-override:true
207+ # concurrency:
208+ # group: ${{ github.workflow }}-vllm-rayserve-cuda-test-${{ github.event.pull_request.number }}
209+ # cancel-in-progress: true
210+ # steps:
211+ # - name: Checkout DLC source
212+ # uses: actions/checkout@v5
213+
214+ # - name: Container pull
215+ # uses: ./.github/actions/ecr-authenticate
216+ # with:
217+ # aws-account-id: ${{ needs.set-rayserve-test-environment.outputs.aws-account-id }}
218+ # aws-region: ${{ vars.AWS_REGION }}
219+ # image-uri: ${{ needs.set-rayserve-test-environment.outputs.image-uri }}
220+
221+ # - name: Checkout vLLM tests
222+ # uses: actions/checkout@v5
223+ # with:
224+ # repository: vllm-project/vllm
225+ # ref: v${{ needs.set-rayserve-test-environment.outputs.framework-version }}
226+ # path: vllm_source
227+
228+ # - name: Start container
229+ # run: |
230+ # CONTAINER_ID=$(docker run -d -it --rm --gpus=all --entrypoint /bin/bash \
231+ # -v ${HOME}/.cache/huggingface:/root/.cache/huggingface \
232+ # -v ${HOME}/.cache/vllm:/root/.cache/vllm \
233+ # -v .:/workdir --workdir /workdir \
234+ # -e HF_TOKEN=${{ secrets.HUGGING_FACE_HUB_TOKEN }} \
235+ # ${{ needs.set-rayserve-test-environment.outputs.image-uri }})
236+ # echo "CONTAINER_ID=$CONTAINER_ID" >> $GITHUB_ENV
237+
238+ # - name: Setup for vLLM tests
239+ # run: |
240+ # docker exec ${CONTAINER_ID} scripts/vllm/vllm_0_10_2_test_setup.sh
241+
242+ # - name: Run vLLM tests
243+ # run: |
244+ # docker exec ${CONTAINER_ID} scripts/vllm/vllm_cuda_test.sh
245+
246+ # vllm-rayserve-example-test:
247+ # needs: [build-vllm-rayserve-image, set-rayserve-test-environment]
248+ # if: success()
249+ # runs-on:
250+ # - codebuild-runner-${{ github.run_id }}-${{ github.run_attempt }}
251+ # fleet:x86-g6xl-runner
252+ # buildspec-override:true
253+ # concurrency:
254+ # group: ${{ github.workflow }}-vllm-rayserve-example-test-${{ github.event.pull_request.number }}
255+ # cancel-in-progress: true
256+ # steps:
257+ # - name: Checkout DLC source
258+ # uses: actions/checkout@v5
259+
260+ # - name: Container pull
261+ # uses: ./.github/actions/ecr-authenticate
262+ # with:
263+ # aws-account-id: ${{ needs.set-rayserve-test-environment.outputs.aws-account-id }}
264+ # aws-region: ${{ vars.AWS_REGION }}
265+ # image-uri: ${{ needs.set-rayserve-test-environment.outputs.image-uri }}
266+
267+ # - name: Checkout vLLM tests
268+ # uses: actions/checkout@v5
269+ # with:
270+ # repository: vllm-project/vllm
271+ # ref: v${{ needs.set-rayserve-test-environment.outputs.framework-version }}
272+ # path: vllm_source
273+
274+ # - name: Start container
275+ # run: |
276+ # CONTAINER_ID=$(docker run -d -it --rm --gpus=all --entrypoint /bin/bash \
277+ # -v ${HOME}/.cache/huggingface:/root/.cache/huggingface \
278+ # -v ${HOME}/.cache/vllm:/root/.cache/vllm \
279+ # -v .:/workdir --workdir /workdir \
280+ # -e HF_TOKEN=${{ secrets.HUGGING_FACE_HUB_TOKEN }} \
281+ # ${{ needs.set-rayserve-test-environment.outputs.image-uri }})
282+ # echo "CONTAINER_ID=$CONTAINER_ID" >> $GITHUB_ENV
283+
284+ # - name: Setup for vLLM tests
285+ # run: |
286+ # docker exec ${CONTAINER_ID} scripts/vllm/vllm_0_10_2_test_setup.sh
287+
288+ # - name: Run vLLM tests
289+ # run: |
290+ # docker exec ${CONTAINER_ID} scripts/vllm/vllm_rayserve_examples_test.sh
291291
292292 generate-rayserve-release-spec :
293- needs : [load-config, build-vllm-rayserve-image, vllm-rayserve-regression-test, vllm-rayserve-cuda-test, vllm-rayserve-example-test]
293+ # needs: [load-config, build-vllm-rayserve-image, vllm-rayserve-regression-test, vllm-rayserve-cuda-test, vllm-rayserve-example-test]
294+ needs : [load-config, build-vllm-rayserve-image]
295+ # if: |
296+ # always() && !failure() && !cancelled() &&
297+ # needs.build-vllm-rayserve-image.result == 'success' &&
298+ # needs.vllm-rayserve-regression-test.result == 'success' &&
299+ # needs.vllm-rayserve-cuda-test.result == 'success' &&
300+ # needs.vllm-rayserve-example-test.result == 'success'
294301 if : |
295302 always() && !failure() && !cancelled() &&
296- needs.build-vllm-rayserve-image.result == 'success' &&
297- needs.vllm-rayserve-regression-test.result == 'success' &&
298- needs.vllm-rayserve-cuda-test.result == 'success' &&
299- needs.vllm-rayserve-example-test.result == 'success'
303+ needs.build-vllm-rayserve-image.result == 'success'
300304 runs-on : ubuntu-latest
301305 outputs :
302306 release-spec : ${{ steps.generate.outputs.release-spec }}
@@ -322,10 +326,8 @@ jobs:
322326
323327 release-rayserve-image :
324328 needs : [load-config, build-vllm-rayserve-image, generate-rayserve-release-spec]
325- # Allow gamma releases from PRs (for testing), but block production releases
326329 if : |
327- needs.generate-rayserve-release-spec.outputs.should-release == 'true' &&
328- (github.event_name != 'pull_request' || fromJson(needs.load-config.outputs.rayserve-config).release.environment == 'gamma')
330+ needs.generate-rayserve-release-spec.outputs.should-release == 'true'
329331 uses : ./.github/workflows/reusable-release-image.yml
330332 with :
331333 source-image-uri : ${{ needs.build-vllm-rayserve-image.outputs.ci-image }}
0 commit comments