This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
d1007767c5f5dd58752f87f39ca4e7847ab69432
vllm
/
tests
/
v1
/
attention
History
Chen Zhang
c75c2e70d6
[Deepseek v3.2] Support indexer prefill chunking (
#25999
)
...
Signed-off-by: Chen Zhang <
zhangch99@outlook.com
> Signed-off-by: simon-mo <
simon.mo@hey.com
>
2025-10-02 10:35:51 -07:00
..
test_attention_backends_selection.py
[Attention] Unify mamba and attention backend selection (
#23171
)
2025-08-25 09:09:36 +00:00
test_attention_backends.py
[V0 deprecation] Remove _VLLM_V1 suffixes from attention backend names (
#25489
)
2025-09-25 17:37:50 +00:00
test_attention_splitting.py
[Bug] fix import and unit test (
#25558
)
2025-09-24 10:17:59 +00:00
test_chunked_local_attention.py
fix some typos (
#24071
)
2025-09-02 20:44:50 -07:00
test_mla_backends.py
[New Model] DeepSeek-V3.2 (Rebased to Main) (
#25896
)
2025-09-30 22:36:24 -07:00
test_sparse_mla_backends.py
[Deepseek v3.2] Support indexer prefill chunking (
#25999
)
2025-10-02 10:35:51 -07:00
utils.py
[New Model] DeepSeek-V3.2 (Rebased to Main) (
#25896
)
2025-09-30 22:36:24 -07:00