Skip to content

Commit 2ce715f

Browse files
committed
Disable Triton FA path in some problematic Paligemma and Phi3v tests
1 parent 39c2cc5 commit 2ce715f

File tree

2 files changed

+12
-1
lines changed

2 files changed

+12
-1
lines changed

tests/distributed/test_multimodal_broadcast.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,14 +15,19 @@
1515

1616
import pytest
1717

18-
from vllm.utils import cuda_device_count_stateless
18+
from vllm.utils import cuda_device_count_stateless, is_hip
1919

2020
model = os.environ["TEST_DIST_MODEL"]
2121

2222
if model.startswith("llava-hf/llava"):
2323
from ..models.test_llava import models, run_test
2424
elif model.startswith("microsoft/Phi-3-vision"):
2525
from ..models.test_phi3v import models, run_test
26+
27+
# ROCm Triton FA runs into issues with these tests, use other backends
28+
# FIXME (mattwong, gshtrasb, hongxiayan)
29+
if is_hip():
30+
os.environ["VLLM_USE_TRITON_FLASH_ATTN"] = "0"
2631
else:
2732
raise NotImplementedError(f"Unsupported model: {model}")
2833

tests/models/test_paligemma.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,4 @@
1+
import os
12
from typing import List, Optional, Tuple, Type
23

34
import pytest
@@ -23,6 +24,11 @@
2324

2425
models = ["google/paligemma-3b-mix-224"]
2526

27+
# ROCm Triton FA runs into issues with these tests, use other backends
28+
# FIXME (mattwong, gshtrasb, hongxiayan)
29+
if is_hip():
30+
os.environ["VLLM_USE_TRITON_FLASH_ATTN"] = "0"
31+
2632

2733
def vllm_to_hf_output(vllm_output: Tuple[List[int], str,
2834
Optional[SampleLogprobs]],

0 commit comments

Comments
 (0)