Skip to content

Pull requests: flashinfer-ai/flashinfer

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Trtllm-gen Fp4 MoE Autotunner
#1475 opened Aug 12, 2025 by IwakuraRein Draft
3 of 5 tasks
Fix TRTLLM NVFP4-out attention kernel scale factor dim issue
#1460 opened Aug 11, 2025 by elvischenv Loading…
4 of 5 tasks
feat(attention): add RoPE offset support for batch prefill
#1457 opened Aug 11, 2025 by MengAiDev Loading…
3 tasks done
Fix cuda-python v13.0 import compatibility
#1455 opened Aug 11, 2025 by yongwww Loading…
3 of 5 tasks
benchmark: add allreduce_fusion benchmark
#1450 opened Aug 10, 2025 by yyihuang Draft
5 tasks
refactor: unify autotuner for fp4 gemm backends
#1439 opened Aug 8, 2025 by ttyio Loading…
3 of 5 tasks
misc: Fix persistent kernel compilation
#1430 opened Aug 8, 2025 by Edenzzzz Loading…
5 tasks
Sink attention AoT
#1427 opened Aug 8, 2025 by nandor Loading…
5 tasks done
Restore llama4 fc2 required kernels
#1417 opened Aug 8, 2025 by aleozlx Loading…
5 tasks done
Removes MPI dependency from MNNVL AllReduce
#1379 opened Aug 4, 2025 by pranavm-nvidia Loading…
5 tasks
feat: Support sliding window for persistent kernel
#1368 opened Aug 3, 2025 by Edenzzzz Loading…
5 tasks
Add moe benchmark routine
#1327 opened Jul 25, 2025 by aleozlx Draft
3 of 5 tasks
Add k_scale and v_scale to persistent attention
#1322 opened Jul 24, 2025 by Edenzzzz Loading…
5 tasks
Wrap cudnn backend to unified interface
#1312 opened Jul 23, 2025 by cyx-6 Loading…
5 tasks
Api regression test for trtllmgen fp8 moe
#1308 opened Jul 23, 2025 by aleozlx Loading…
5 tasks done
fix: a workaround to make fp8 kv-cache work for prefill
#1304 opened Jul 22, 2025 by chenyang78 Loading…
2 tasks
ProTip! Type g p on any issue or pull request to go back to the pull request listing page.