Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Support AMD ROCm on FlashAttention 2
#1010 opened Jun 26, 2024 by rocking5566 Loading…
Add return_softmax_lse in flash_attn_with_kvcache
#989 opened Jun 13, 2024 by ovowei Loading…
Fix +/-inf in LSE returned by forward
#978 opened Jun 3, 2024 by sgrigory Loading…
add pyproject.toml with build dependencies
#958 opened May 17, 2024 by dhellmann Loading…
Relative position encoding
#956 opened May 14, 2024 by b-albar Loading…
1 of 4 tasks
Add softmax_d in mha_bwd
#905 opened Apr 1, 2024 by MayDomine Loading…
ALiBi for the non-flash code path
#858 opened Feb 29, 2024 by Markus28 Loading…
Add support for small page sizes
#824 opened Feb 13, 2024 by skrider Loading…
Add C++ build support for use with LibTorch
#819 opened Feb 9, 2024 by shaltielshmid Loading…
meta tensor stuff
#769 opened Jan 15, 2024 by tsengalb99 Loading…
Jetson (aarch64) support
#724 opened Dec 14, 2023 by jasl Loading…
Update utils.py
#710 opened Dec 8, 2023 by adarshxs Loading…
Add flash_attn_varlen_func_with_kvcache.
#685 opened Nov 22, 2023 by garrett4wade Loading…
Custom attention bias
#617 opened Oct 19, 2023 by b-albar Loading…
2 of 5 tasks
Setup: Add extra compute targets
#605 opened Oct 15, 2023 by bdashore3 Loading…
Support returning attention weights in naive attention modules
#589 opened Oct 4, 2023 by kklemon Loading…
1 task
A100 FAv2 not working
#572 opened Sep 26, 2023 by ksivaman Loading…
[Rotary] more varlen rotary function implement
#556 opened Sep 19, 2023 by GGGGGGXY Loading…
Support key padding masks in GPT models
#544 opened Sep 14, 2023 by kevinhu Loading…
skip git commands when not inside a git repo
#515 opened Sep 5, 2023 by tmm1 Loading…
ProTip! no:milestone will show everything without a milestone.