We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 894783c commit ac9c7f3Copy full SHA for ac9c7f3
src/transformers/models/llama4/modeling_llama4.py
@@ -490,7 +490,7 @@ class Llama4PreTrainedModel(PreTrainedModel):
490
config_class = Llama4Config
491
supports_gradient_checkpointing = True
492
_skip_keys_device_placement = ["past_key_values"]
493
- _supports_flash_attn_2 = True
+ _supports_flash_attn_2 = False
494
_supports_sdpa = True
495
_supports_flex_attn = True
496
_supports_cache_class = True
0 commit comments