Skip to content

Commit 758b239

Browse files
authored
Refine exllamav2 ut (#1001)
1 parent c0aa30a commit 758b239

File tree

1 file changed

+4
-2
lines changed

1 file changed

+4
-2
lines changed

test/test_cuda/test_exllamav2_backend.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212

1313
from auto_round import AutoRound, AutoRoundConfig
1414
from auto_round.eval.evaluation import simple_evaluate_user_model
15-
from auto_round.testing_utils import require_autogptq, require_gptqmodel
15+
from auto_round.testing_utils import require_autogptq, require_gptqmodel, require_package_version_ut
1616

1717

1818
class LLMDataLoader:
@@ -24,7 +24,7 @@ def __iter__(self):
2424
yield torch.ones([1, 10], dtype=torch.long)
2525

2626

27-
class TestAutoRoundMarlinBackend(unittest.TestCase):
27+
class TestAutoRoundexllamaBackend(unittest.TestCase):
2828

2929
@classmethod
3030
def setUpClass(self):
@@ -99,6 +99,7 @@ def test_gptqmodel_exllmav2_4bits_asym(self):
9999
shutil.rmtree("./saved", ignore_errors=True)
100100

101101
@require_autogptq
102+
@require_package_version_ut("torch", "<2.6.0")
102103
def test_gptq_exllamav2_4bits_sym(self):
103104
model = AutoModelForCausalLM.from_pretrained(self.model_name, torch_dtype="auto", trust_remote_code=True)
104105
tokenizer = AutoTokenizer.from_pretrained(self.model_name, trust_remote_code=True)
@@ -130,6 +131,7 @@ def test_gptq_exllamav2_4bits_sym(self):
130131
shutil.rmtree(self.save_folder, ignore_errors=True)
131132

132133
@require_autogptq
134+
@require_package_version_ut("torch", "<2.6.0")
133135
def test_gptq_exllamav2_4bits_sym_group_size(self):
134136
for group_size in [-1, 32, 64, 128, 256, 1024]: ## 384, 768 has accuracy issue
135137
print(f"!!!!!!!!!!!!!!!!!{group_size}!!!!!!!!!!!!!!!!!")

0 commit comments

Comments
 (0)