This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
a28b94e6ef60b7f5aa1b97bc8d966a8d12cbc1da
vllm
/
vllm
/
v1
/
attention
/
ops
History
Lucas Wilkinson
889722f3bf
[FlashMLA] Update FlashMLA to expose new arguments (
#32810
)
...
Signed-off-by: Lucas Wilkinson <
lwilkins@redhat.com
>
2026-01-21 22:02:39 -07:00
..
__init__.py
…
chunked_prefill_paged_decode.py
…
common.py
…
flashmla.py
[FlashMLA] Update FlashMLA to expose new arguments (
#32810
)
2026-01-21 22:02:39 -07:00
merge_attn_states.py
…
paged_attn.py
[2/N][Attention] Fix pre-commit errors (
#32052
)
2026-01-10 00:27:15 +00:00
prefix_prefill.py
…
rocm_aiter_mla_sparse.py
[ROCm][Deepseekv3.2] Refactor Sparse Indexer as CustomOp (
#29287
)
2026-01-21 23:16:30 +08:00
triton_decode_attention.py
…
triton_merge_attn_states.py
…
triton_prefill_attention.py
[Performance] Improve Triton prefill attention kernel's performance (
#32403
)
2026-01-17 20:19:59 -08:00
triton_reshape_and_cache_flash.py
…
triton_unified_attention.py
[Model] Support Step1 Model (
#32511
)
2026-01-18 10:20:46 +00:00
vit_attn_wrappers.py
[Misc][LLaMa4] Compile LLaMa Vision Encoder (
#30709
)
2026-01-09 22:01:38 -05:00