Skip to content

Commit e8334d9

Browse files
[ROCm][CI] Keeping AITER FA attention for whisper pending vllm-project#28376
Signed-off-by: Andreas Karatzas <akaratza@amd.com>
1 parent aa2a7f7 commit e8334d9

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

tests/entrypoints/openai/test_transcription_validation.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -28,18 +28,18 @@
2828

2929

3030
@pytest.fixture(scope="module", autouse=True)
31-
def rocm_flex_attention():
31+
def rocm_aiter_fa_attention():
3232
"""
33-
Sets VLLM_ATTENTION_BACKEND=FLEX_ATTENTION for ROCm
34-
for the duration of this test module. For now the only
35-
attention backend that supports cross attention on ROCm
36-
is FLEX_ATTENTION.
33+
Sets VLLM_ATTENTION_BACKEND=ROCM_AITER_FA for ROCm
34+
for the duration of this test module. #28376 introduces
35+
support for encoder_decoder attention tested on
36+
WhisperForConditionalGeneration architecture.
3737
"""
3838
from vllm.platforms import current_platform
3939

4040
if current_platform.is_rocm():
4141
old_backend = os.environ.get("VLLM_ATTENTION_BACKEND")
42-
os.environ["VLLM_ATTENTION_BACKEND"] = "FLEX_ATTENTION"
42+
os.environ["VLLM_ATTENTION_BACKEND"] = "ROCM_AITER_FA"
4343
yield
4444
if old_backend is None:
4545
del os.environ["VLLM_ATTENTION_BACKEND"]

0 commit comments

Comments
 (0)