Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

feat: add support for torch2.7
#1574 opened Apr 3, 2025 by NanoCode012 Draft
[BugFix] Ensure tensor contiguity.
#1562 opened Mar 28, 2025 by ibganev Loading…
Add compile support for flash_attn_with_kvcache
#1554 opened Mar 25, 2025 by ani300 Loading…
Support cuda 12.8.0 and SBSA wheels
#1507 opened Feb 25, 2025 by johnnynunez Loading…
[docs][FA3] Update build guide of FlashAttention-3
#1488 opened Feb 11, 2025 by Ethkuil Loading…
Remove ninja runtime dependency
#1484 opened Feb 9, 2025 by kevmo314 Loading…
Add missing tests/__init__.py
#1405 opened Dec 20, 2024 by BioGeek Loading…
Fix deprecation warnings
#1382 opened Dec 12, 2024 by rongou Loading…
flashattnvarlen support tree attention
#1188 opened Aug 30, 2024 by efsotr Loading…
add softmax_d for mha_bwd
#1161 opened Aug 19, 2024 by MayDomine Loading…
Windows actions
#1036 opened Jul 9, 2024 by kingbri1 Loading…
change condition to num_heads >= num_heads_k
#1030 opened Jul 5, 2024 by xenshinu Loading…
Fix +/-inf in LSE returned by forward
#978 opened Jun 3, 2024 by sgrigory Loading…
add pyproject.toml with build dependencies
#958 opened May 17, 2024 by dhellmann Loading…
Relative position encoding
#956 opened May 14, 2024 by b-albar Loading…
1 of 4 tasks
ProTip! Follow long discussions with comments:>50.