Skip to content

Commit d65668b

Browse files
Concurrenseegemini-code-assist[bot]DarkLight1337
authored
Adding "AMD: Multi-step Tests" to amdproduction. (vllm-project#19508)
Signed-off-by: Yida Wu <[email protected]> Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> Co-authored-by: Cyrus Leung <[email protected]>
1 parent aafbbd9 commit d65668b

File tree

2 files changed

+11
-1
lines changed

2 files changed

+11
-1
lines changed

.buildkite/test-pipeline.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -675,7 +675,7 @@ steps:
675675
- pytest -v -s plugins/lora_resolvers # unit tests for in-tree lora resolver plugins
676676

677677
- label: Multi-step Tests (4 GPUs) # 36min
678-
mirror_hardwares: [amdexperimental]
678+
mirror_hardwares: [amdexperimental, amdproduction]
679679
working_dir: "/vllm-workspace/tests"
680680
num_gpus: 4
681681
source_file_dependencies:

tests/multi_step/test_correctness_llm.py

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,7 @@
88

99
import pytest
1010

11+
from vllm.platforms import current_platform
1112
from vllm.utils import STR_BACKEND_ENV_VAR
1213

1314
from ..models.utils import check_logprobs_close, check_outputs_equal
@@ -71,6 +72,12 @@ def test_multi_step_llm(
7172
num_logprobs: corresponds to the `logprobs` argument to the OpenAI
7273
completions endpoint; `None` -> 1 logprob returned.
7374
"""
75+
if current_platform.is_rocm() and \
76+
(attention_backend == "FLASHINFER" or enable_chunked_prefill):
77+
pytest.skip(
78+
"Multi-Step with FLASHINFER or Chunked-Prefill is not supported"
79+
"on ROCm")
80+
7481
with monkeypatch.context() as m:
7582
m.setenv(STR_BACKEND_ENV_VAR, attention_backend)
7683

@@ -221,6 +228,9 @@ def test_multi_step_llm_w_prompt_logprobs(
221228
@pytest.mark.parametrize("num_prompts", NUM_PROMPTS)
222229
@pytest.mark.parametrize("num_logprobs", [None, 5])
223230
@pytest.mark.parametrize("attention_backend", ["FLASH_ATTN"])
231+
@pytest.mark.skipif(
232+
current_platform.is_rocm(),
233+
reason="Multi-Step + Chunked-Prefill not supported on ROCm")
224234
def test_multi_step_llm_chunked_prefill_prefix_cache(
225235
vllm_runner,
226236
example_prompts,

0 commit comments

Comments
 (0)