forked from Dao-AILab/flash-attention
-
Notifications
You must be signed in to change notification settings - Fork 108
Pull requests: vllm-project/flash-attention
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[CI/Build] make it possible to build with a free-threaded interpreter
#112
opened Nov 22, 2025 by
rgommers
Loading…
Add num_splits for mha_varlen_fwd FA2, support batch invariant
#110
opened Nov 17, 2025 by
quanliu1991
Loading…
[Kernel] add attention sinks for flash attention2
#103
opened Oct 19, 2025 by
dudugong-gitch
Loading…
feat: implement tree attention mask support for FlashAttention-2
#81
opened Aug 15, 2025 by
foolusion
Loading…
Removed the assertion imposed on cu_seqlens_k and seqused_k
#59
opened Mar 29, 2025 by
chenyang78
Loading…
Add back flash_attn_func api (and support FA3) [Don't Merge Yet]
#40
opened Jan 26, 2025 by
LucasWilkinson
Loading…
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.