Skip to content

Commit

Permalink
skip rocm
Browse files Browse the repository at this point in the history
  • Loading branch information
Varun Sundar Rabindranath committed Oct 6, 2024
1 parent a1cb996 commit 6cad135
Showing 1 changed file with 4 additions and 6 deletions.
10 changes: 4 additions & 6 deletions tests/core/test_num_computed_tokens_update.py
Original file line number Diff line number Diff line change
@@ -1,10 +1,9 @@
import os

import pytest

from tests.conftest import VllmRunner
from tests.core.utils import create_dummy_prompt
from vllm.engine.llm_engine import LLMEngine
from vllm.platforms import current_platform
from vllm.sequence import SequenceGroup

MODEL = "JackFram/llama-160m"
Expand All @@ -25,10 +24,9 @@ def test_num_computed_tokens_update(num_scheduler_steps: int,
is_multi_step = num_scheduler_steps > 1
is_multi_step_chunked_prefill = is_multi_step and enable_chunked_prefill

attention_backend = os.getenv("VLLM_ATTENTION_BACKEND", "FLASH_ATTN")
if is_multi_step_chunked_prefill and attention_backend != "FLASH_ATTN":
pytest.skip("Multi-step with Chunked-Prefill only supports"
" FLASH_ATTN backend")
if is_multi_step_chunked_prefill and current_platform.is_rocm():
pytest.skip("Multi-step with Chunked-Prefill does not support "
"rocm_flash_attn backend")

# Make a vllm engine
runner = VllmRunner(model_name=MODEL,
Expand Down

0 comments on commit 6cad135

Please sign in to comment.