Skip to content

Pull requests: PaddlePaddle/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Add FA4 varlen.
#129 opened Apr 13, 2026 by baoqiwen Loading…
bwd support (192, 128) for sm100
#123 opened Apr 3, 2026 by baoqiwen Loading…
Tune registers
#122 opened Apr 1, 2026 by baoqiwen Loading…
FlashMaskV3 Single-node Speed Optimization
#119 opened Mar 26, 2026 by Enigmatisms Loading…
Add rrattn estimate func and interface
#117 opened Mar 13, 2026 by LLSGYN Loading…
Support Global Sliding Window (num_vec == 4) on FM4 BWD
#111 opened Mar 3, 2026 by umiswing Member Loading…
adapt to torch version flashmaskv4
#103 opened Jan 28, 2026 by clouds1238 Loading…
add_flashmask_cpbalance
#99 opened Dec 30, 2025 by starcrown001 Loading…
fine-tuned tile size & regitser for fwd_hdim64
#92 opened Nov 14, 2025 by xxyux Loading…
fix fa2 flashmask oob read
#67 opened Jun 26, 2025 by umiswing Member Draft
[WIP] fa3 varlen fix int32 overflow
#65 opened Jun 19, 2025 by umiswing Member Loading…
optimize skip block calculate in bwd
#49 opened Aug 28, 2024 by GuoxiaWang Collaborator Loading…
[BugFix] fix_mask error using unpadding api
#41 opened Apr 23, 2024 by wwbitejotunn Loading…
Fa cmake extends op
#31 opened Dec 14, 2023 by AnnaTrainingG Loading…
Fa cmake
#29 opened Dec 6, 2023 by AnnaTrainingG Loading…
[WIP]Sparse seqparallel
#9 opened Jun 8, 2023 by zkh2016 Loading…
add block sparse api
#7 opened May 27, 2023 by kuizhiqing Member Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.