We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 76879cc commit f643ec2Copy full SHA for f643ec2
vllm/platforms/rocm.py
@@ -276,6 +276,9 @@ def get_attn_backend_cls(
276
)
277
278
if envs.VLLM_USE_V1:
279
+ if selected_backend == _Backend.FLEX_ATTENTION:
280
+ logger.info("Using FlexAttention backend on V1 engine.")
281
+ return "vllm.v1.attention.backends.flex_attention.FlexAttentionBackend"
282
if (
283
envs.VLLM_ROCM_USE_AITER and envs.VLLM_ROCM_USE_AITER_MHA and on_gfx9()
284
) or selected_backend == _Backend.ROCM_AITER_FA:
0 commit comments