This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
2c0ee0fde8db447bd6e7d3ea0d49070661a243fa
vllm
/
tests
/
models
/
multimodal
History
Isotr0py
55f1fc1b1b
[v1] Add PrefixLM support to TritonAttention backend (
#30386
)
...
(cherry picked from commit
74a1ac38b0
)
2025-12-17 19:57:52 -08:00
..
generation
[v1] Add PrefixLM support to TritonAttention backend (
#30386
)
2025-12-17 19:57:52 -08:00
pooling
Support tokenization_kwargs override (
#29794
)
2025-12-06 09:12:53 +00:00
processing
[Core][MM] Optimize encoder cache manager by operating with embeddings only (
#30475
)
2025-12-16 17:15:49 -08:00
__init__.py
[CI/Build] Move model-specific multi-modal processing tests (
#11934
)
2025-01-11 13:50:05 +08:00
test_mapping.py
Revert "[Renderer] Separate out
RendererConfig
from
ModelConfig
(
#30145
)" (
#30199
)
2025-12-07 00:00:22 -08:00