@@ -227,33 +227,33 @@ def test_vlm_gguf(self):
227227 self .assertAlmostEqual (file_size , 1599 , delta = 1.0 )
228228 shutil .rmtree (quantized_model_path , ignore_errors = True )
229229
230- @require_gguf
231- def test_llama_4 (self ):
232- model_name = "/dataset/Llama-4-Scout-17B-16E-Instruct/"
233- from auto_round import AutoRoundMLLM
234- from auto_round .utils import mllm_load_model
235-
236- model , processor , tokenizer , image_processor = mllm_load_model (model_name , use_auto_mapping = False )
237- autoround = AutoRoundMLLM (
238- model ,
239- tokenizer = tokenizer ,
240- processor = processor ,
241- image_processor = image_processor ,
242- device = "auto" ,
243- iters = 0 ,
244- )
245- quantized_model_path = "/dataset/Llam-4-test"
246- shutil .rmtree (quantized_model_path , ignore_errors = True )
247- autoround .quantize_and_save (output_dir = quantized_model_path , format = "gguf:q4_0" )
248- self .assertTrue ("mmproj-model.gguf" in os .listdir (quantized_model_path ))
249- file_size = (
250- os .path .getsize (os .path .join (quantized_model_path , "Llama-4-Scout-17B-16E-Instruct-16x17B-Q4_0.gguf" ))
251- / 1024 ** 2
252- )
253- self .assertAlmostEqual (file_size , 58093.62 , delta = 1.0 )
254- file_size = os .path .getsize (os .path .join (quantized_model_path , "mmproj-model.gguf" )) / 1024 ** 2
255- self .assertAlmostEqual (file_size , 3326.18 , delta = 5.0 )
256- shutil .rmtree (quantized_model_path , ignore_errors = True )
230+ # @require_gguf
231+ # def test_llama_4(self):
232+ # model_name = "/dataset/Llama-4-Scout-17B-16E-Instruct/"
233+ # from auto_round import AutoRoundMLLM
234+ # from auto_round.utils import mllm_load_model
235+
236+ # model, processor, tokenizer, image_processor = mllm_load_model(model_name, use_auto_mapping=False)
237+ # autoround = AutoRoundMLLM(
238+ # model,
239+ # tokenizer=tokenizer,
240+ # processor=processor,
241+ # image_processor=image_processor,
242+ # device="auto",
243+ # iters=0,
244+ # )
245+ # quantized_model_path = "/dataset/Llam-4-test"
246+ # shutil.rmtree(quantized_model_path, ignore_errors=True)
247+ # autoround.quantize_and_save(output_dir=quantized_model_path, format="gguf:q4_0")
248+ # self.assertTrue("mmproj-model.gguf" in os.listdir(quantized_model_path))
249+ # file_size = (
250+ # os.path.getsize(os.path.join(quantized_model_path, "Llama-4-Scout-17B-16E-Instruct-16x17B-Q4_0.gguf"))
251+ # / 1024**2
252+ # )
253+ # self.assertAlmostEqual(file_size, 58093.62, delta=1.0)
254+ # file_size = os.path.getsize(os.path.join(quantized_model_path, "mmproj-model.gguf")) / 1024**2
255+ # self.assertAlmostEqual(file_size, 3326.18, delta=5.0)
256+ # shutil.rmtree(quantized_model_path, ignore_errors=True)
257257
258258
259259if __name__ == "__main__" :
0 commit comments