1212
1313from auto_round import AutoRound , AutoRoundConfig
1414from auto_round .eval .evaluation import simple_evaluate_user_model
15- from auto_round .testing_utils import require_autogptq , require_gptqmodel
15+ from auto_round .testing_utils import require_autogptq , require_gptqmodel , require_package_version_ut
1616
1717
1818class LLMDataLoader :
@@ -24,7 +24,7 @@ def __iter__(self):
2424 yield torch .ones ([1 , 10 ], dtype = torch .long )
2525
2626
27- class TestAutoRoundMarlinBackend (unittest .TestCase ):
27+ class TestAutoRoundexllamaBackend (unittest .TestCase ):
2828
2929 @classmethod
3030 def setUpClass (self ):
@@ -99,6 +99,7 @@ def test_gptqmodel_exllmav2_4bits_asym(self):
9999 shutil .rmtree ("./saved" , ignore_errors = True )
100100
101101 @require_autogptq
102+ @require_package_version_ut ("torch" , "<2.6.0" )
102103 def test_gptq_exllamav2_4bits_sym (self ):
103104 model = AutoModelForCausalLM .from_pretrained (self .model_name , torch_dtype = "auto" , trust_remote_code = True )
104105 tokenizer = AutoTokenizer .from_pretrained (self .model_name , trust_remote_code = True )
@@ -130,6 +131,7 @@ def test_gptq_exllamav2_4bits_sym(self):
130131 shutil .rmtree (self .save_folder , ignore_errors = True )
131132
132133 @require_autogptq
134+ @require_package_version_ut ("torch" , "<2.6.0" )
133135 def test_gptq_exllamav2_4bits_sym_group_size (self ):
134136 for group_size in [- 1 , 32 , 64 , 128 , 256 , 1024 ]: ## 384, 768 has accuracy issue
135137 print (f"!!!!!!!!!!!!!!!!!{ group_size } !!!!!!!!!!!!!!!!!" )
0 commit comments