2525from transformers .testing_utils import (
2626 cleanup ,
2727 require_bitsandbytes ,
28+ require_read_token ,
2829 require_torch ,
2930 require_torch_gpu ,
3031 slow ,
@@ -315,6 +316,7 @@ def setUp(self):
315316 def tearDown (self ):
316317 cleanup (torch_device , gc_collect = True )
317318
319+ @require_read_token
318320 def test_mistral3_integration_generate_text_only (self ):
319321 processor = AutoProcessor .from_pretrained (self .model_checkpoint )
320322 model = Mistral3ForConditionalGeneration .from_pretrained (
@@ -342,6 +344,7 @@ def test_mistral3_integration_generate_text_only(self):
342344 expected_output = "Sure, here's a haiku for you:\n \n Whispers of the breeze,\n Cherry blossoms softly fall,\n Spring's gentle embrace."
343345 self .assertEqual (decoded_output , expected_output )
344346
347+ @require_read_token
345348 def test_mistral3_integration_generate (self ):
346349 processor = AutoProcessor .from_pretrained (self .model_checkpoint )
347350 model = Mistral3ForConditionalGeneration .from_pretrained (
@@ -368,6 +371,7 @@ def test_mistral3_integration_generate(self):
368371 expected_output = "The image depicts two cats lying on a pink blanket. The larger cat, which appears to be an"
369372 self .assertEqual (decoded_output , expected_output )
370373
374+ @require_read_token
371375 def test_mistral3_integration_batched_generate (self ):
372376 processor = AutoProcessor .from_pretrained (self .model_checkpoint )
373377 model = Mistral3ForConditionalGeneration .from_pretrained (
@@ -418,6 +422,7 @@ def test_mistral3_integration_batched_generate(self):
418422 f"Decoded output: { decoded_output } \n Expected output: { expected_output } " ,
419423 )
420424
425+ @require_read_token
421426 @require_bitsandbytes
422427 def test_mistral3_integration_batched_generate_multi_image (self ):
423428 processor = AutoProcessor .from_pretrained (self .model_checkpoint )
0 commit comments