We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 0dabbf1 commit 46284b7Copy full SHA for 46284b7
vllm/model_executor/models/qwen2_5_vl.py
@@ -424,6 +424,7 @@ def forward(
424
elif self.attn_backend == _Backend.TORCH_SDPA:
425
# Execute attention entry by entry for speed & less VRAM.
426
from vllm.platforms import current_platform
427
+
428
if current_platform.is_rocm():
429
q = q.contiguous()
430
k = k.contiguous()
vllm/model_executor/models/qwen2_vl.py
@@ -454,6 +454,7 @@ def forward(
454
455
456
457
458
459
460
0 commit comments