File tree Expand file tree Collapse file tree 2 files changed +11
-1
lines changed
Expand file tree Collapse file tree 2 files changed +11
-1
lines changed Original file line number Diff line number Diff line change @@ -675,7 +675,7 @@ steps:
675675 - pytest -v -s plugins/lora_resolvers # unit tests for in-tree lora resolver plugins
676676
677677- label : Multi-step Tests (4 GPUs) # 36min
678- mirror_hardwares : [amdexperimental]
678+ mirror_hardwares : [amdexperimental, amdproduction ]
679679 working_dir : " /vllm-workspace/tests"
680680 num_gpus : 4
681681 source_file_dependencies :
Original file line number Diff line number Diff line change 88
99import pytest
1010
11+ from vllm .platforms import current_platform
1112from vllm .utils import STR_BACKEND_ENV_VAR
1213
1314from ..models .utils import check_logprobs_close , check_outputs_equal
@@ -71,6 +72,12 @@ def test_multi_step_llm(
7172 num_logprobs: corresponds to the `logprobs` argument to the OpenAI
7273 completions endpoint; `None` -> 1 logprob returned.
7374 """
75+ if current_platform .is_rocm () and \
76+ (attention_backend == "FLASHINFER" or enable_chunked_prefill ):
77+ pytest .skip (
78+ "Multi-Step with FLASHINFER or Chunked-Prefill is not supported"
79+ "on ROCm" )
80+
7481 with monkeypatch .context () as m :
7582 m .setenv (STR_BACKEND_ENV_VAR , attention_backend )
7683
@@ -221,6 +228,9 @@ def test_multi_step_llm_w_prompt_logprobs(
221228@pytest .mark .parametrize ("num_prompts" , NUM_PROMPTS )
222229@pytest .mark .parametrize ("num_logprobs" , [None , 5 ])
223230@pytest .mark .parametrize ("attention_backend" , ["FLASH_ATTN" ])
231+ @pytest .mark .skipif (
232+ current_platform .is_rocm (),
233+ reason = "Multi-Step + Chunked-Prefill not supported on ROCm" )
224234def test_multi_step_llm_chunked_prefill_prefix_cache (
225235 vllm_runner ,
226236 example_prompts ,
You can’t perform that action at this time.
0 commit comments