-
Notifications
You must be signed in to change notification settings - Fork 561
Pull requests: flashinfer-ai/flashinfer
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Fix: several bugs/issues with trtllm-gen attention kernels.
#2062
opened Nov 7, 2025 by
PerkzZheng
Loading…
5 tasks
Rebase FP8 SM100 Cutlass FMHA Attention to main (original PR#1238)
#2047
opened Nov 5, 2025 by
pavanimajety
•
Draft
5 tasks
feat: Add flashinfer.rope.rope_quantize_fp8_append_paged_kv_cache (fused RoPE + Q + KV cache, supports MLA/GQA/MHA)
#2037
opened Nov 4, 2025 by
kahyunnam
Loading…
5 tasks done
Enable renormalize(naive) routing for fp8 per-tensor
#2030
opened Nov 3, 2025 by
IwakuraRein
•
Draft
5 tasks
Refactor flashinfer/__init__.py so that applications could selectively pack submodules without modifying __init__.py
#2027
opened Nov 3, 2025 by
bangshengtang
Loading…
5 tasks done
refactor: backend_requirement + supported_compute_capability decorator for gemm
#2000
opened Oct 29, 2025 by
jimmyzho
Loading…
5 tasks
chore: agentic workflow for automatic version bump
#1947
opened Oct 19, 2025 by
yzh119
Loading…
5 tasks
Fix "cannot find -lcuda & -lcudart" problem in WSL2
#1909
opened Oct 10, 2025 by
HelloCard
Loading…
3 tasks
[DO NOT MERGE][WIP] lint: Add clang-tidy to pre-commits
#1845
opened Oct 2, 2025 by
yzh119
Loading…
5 tasks
chore: allow custom paths for external dependencies like CUTLASS
#1827
opened Oct 1, 2025 by
yzh119
Loading…
4 of 5 tasks
fix the dequantize_block in the trtllm_cutlass fuse moe test
#1721
opened Sep 18, 2025 by
rainj-me
Loading…
5 tasks done
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.