-
Notifications
You must be signed in to change notification settings - Fork 1.9k
Pull requests: Dao-AILab/flash-attention
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[BugFix] Fix flash_attn_with_kvcache with scalar cache_seqlen
#1795
opened Aug 5, 2025 by
stepinto
Loading…
fix race condition bug in cute _flash_attn_fwd in multiple gpu env
#1793
opened Aug 1, 2025 by
beiw-nv
Loading…
Add torch.compile support to flash attention 3
#1769
opened Jul 22, 2025 by
guilhermeleobas
Loading…
Enable the deterministic mode option in the backward kernel
#1766
opened Jul 21, 2025 by
GD06
Loading…
Fix illegal memory access through off-by-one error in num_splits_dynamic_ptr init
#1747
opened Jul 10, 2025 by
klondenberg-bioptimus
Loading…
Useuful command to install flash faster on behamoth clusters
#1660
opened May 10, 2025 by
sleepingcat4
Loading…
Patch
RPATH
of compiled Linux library to locate PyTorch and CUDA libraries in virtual env
#1634
opened Apr 30, 2025 by
sisp
Loading…
feat: support to tile K and V separately in FA3 backward
#1626
opened Apr 28, 2025 by
beginlner
Loading…
add checks for zero elements input of triton LayerNorm impl
#1621
opened Apr 27, 2025 by
Luciennnnnnn
Loading…
Add PT compileable support for flash_attn_with_kvcache
#1592
opened Apr 14, 2025 by
jataylo
Loading…
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.