Skip to content

Merge branch 'main' into codex/transformers-44380-gpt2-sdpa-scaling

81269ef
Select commit
Loading
Failed to load commit list.
Merged

Fix GPT2 attention scaling ignored in SDPA/FlashAttention #44397

Merge branch 'main' into codex/transformers-44380-gpt2-sdpa-scaling
81269ef
Select commit
Loading
Failed to load commit list.