This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
e5b807607c8493155e6eccd665772d4c19b2114e
vllm
/
benchmarks
/
attention_benchmarks
/
configs
History
Matthew Bonanni
f444c05c32
[Attention] Use FA4 for MLA prefill (
#34732
)
...
Signed-off-by: Matthew Bonanni <
mbonanni@redhat.com
>
2026-03-12 12:10:17 -04:00
..
mla_decode.yaml
[Attention] Add FlashInfer Sparse MLA backend (
#33451
)
2026-02-12 17:21:54 +00:00
mla_mixed_batch.yaml
[Attention] Add FlashInfer Sparse MLA backend (
#33451
)
2026-02-12 17:21:54 +00:00
mla_prefill.yaml
[Attention] Use FA4 for MLA prefill (
#34732
)
2026-03-12 12:10:17 -04:00
mla_sparse_prefill.yaml
[Attention] Use FA4 for MLA prefill (
#34732
)
2026-03-12 12:10:17 -04:00
reorder_threshold.yaml
[Attention] Add FlashInfer Sparse MLA backend (
#33451
)
2026-02-12 17:21:54 +00:00
speculative_decode.yaml
[Attention] Add FlashInfer Sparse MLA backend (
#33451
)
2026-02-12 17:21:54 +00:00
standard_attention.yaml
[Attention] Add FlashInfer Sparse MLA backend (
#33451
)
2026-02-12 17:21:54 +00:00