-
Notifications
You must be signed in to change notification settings - Fork 514
Pull requests: flashinfer-ai/flashinfer
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
cleanup: ensure repository URL has no trailing slash
#1759
opened Sep 23, 2025 by
tarukumar
Loading…
3 of 5 tasks
[draft]Fix sink attention accuracy regression, add more sink attention test.
#1758
opened Sep 23, 2025 by
weireweire
Loading…
5 tasks
fix: should pass global_override_indptr_cpu in fast_decode_plan param list
#1757
opened Sep 23, 2025 by
yyihuang
Loading…
5 tasks
feat: Add compute capability checks to flashinfer_benchmark
#1756
opened Sep 23, 2025 by
bkryu
Loading…
5 tasks done
Fix tests/test_trtllm_gen_attention.py::test_trtllm_batch_prefill, ::test_trtllm_batch_decode mismatch error
#1755
opened Sep 22, 2025 by
kahyunnam
Loading…
bugfix: partially fix tests/test_trtllm_gen_fused_moe.py unite test failure
#1724
opened Sep 19, 2025 by
nv-yunzheq
Loading…
5 tasks
fix the dequantize_block in the trtllm_cutlass fuse moe test
#1721
opened Sep 18, 2025 by
rainj-me
Loading…
5 tasks done
fix: add _check_tensor_params to check correct sampling parameters and dtype validation in decode.py
#1652
opened Sep 8, 2025 by
raayandhar
Loading…
5 tasks done
[WIP] refactor: using tvm-ffi for multi-platform bindings
#1641
opened Sep 5, 2025 by
yzh119
Loading…
5 tasks
feat: support radix-based top-k sampling algorithm
#1561
opened Aug 24, 2025 by
JasonJ2021
Loading…
4 of 5 tasks
tests(attn): add short-seq CUDA edge-case test (qo_len=1) for prefill
#1515
opened Aug 19, 2025 by
PrithviElancherran
Loading…
3 tasks done
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.