File tree Expand file tree Collapse file tree 1 file changed +6
-6
lines changed Expand file tree Collapse file tree 1 file changed +6
-6
lines changed Original file line number Diff line number Diff line change 2828
2929
3030@pytest .fixture (scope = "module" , autouse = True )
31- def rocm_flex_attention ():
31+ def rocm_aiter_fa_attention ():
3232 """
33- Sets VLLM_ATTENTION_BACKEND=FLEX_ATTENTION for ROCm
34- for the duration of this test module. For now the only
35- attention backend that supports cross attention on ROCm
36- is FLEX_ATTENTION .
33+ Sets VLLM_ATTENTION_BACKEND=ROCM_AITER_FA for ROCm
34+ for the duration of this test module. #28376 introduces
35+ support for encoder_decoder attention tested on
36+ WhisperForConditionalGeneration architecture .
3737 """
3838 from vllm .platforms import current_platform
3939
4040 if current_platform .is_rocm ():
4141 old_backend = os .environ .get ("VLLM_ATTENTION_BACKEND" )
42- os .environ ["VLLM_ATTENTION_BACKEND" ] = "FLEX_ATTENTION "
42+ os .environ ["VLLM_ATTENTION_BACKEND" ] = "ROCM_AITER_FA "
4343 yield
4444 if old_backend is None :
4545 del os .environ ["VLLM_ATTENTION_BACKEND" ]
You can’t perform that action at this time.
0 commit comments