This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
fa6a6be51978bd4b49ba0da17039e60f96dc5b13
vllm
/
examples
/
pooling
History
Jakub Zakrzewski
111d869069
[Model] Add nvidia/llama-nemotron-embed-vl-1b-v2 multimodal embedding model (
#35297
)
...
Signed-off-by: Jakub Zakrzewski <
jzakrzewski@nvidia.com
>
2026-02-26 14:17:17 +00:00
..
classify
[Doc] Update usage of
--limit-mm-per-prompt
(
#34148
)
2026-02-09 21:12:13 -08:00
embed
[Model] Add nvidia/llama-nemotron-embed-vl-1b-v2 multimodal embedding model (
#35297
)
2026-02-26 14:17:17 +00:00
plugin
(bugfix): Fixed encode in LLM entrypoint for IOProcessr plugin prompts (
#34618
)
2026-02-16 07:33:55 -08:00
pooling
[Frontend][last/5] Make pooling entrypoints request schema consensus. (
#31127
)
2026-02-09 06:42:38 +00:00
score
[New Model] Add ColModernVBERT (
#34558
)
2026-02-22 12:23:41 +08:00
token_classify
[Frontend][2/n] Make pooling entrypoints request schema consensus | ChatRequest (
#32574
)
2026-01-22 10:32:44 +00:00
token_embed
[new model] add COLQwen3 code & Inference (
#34398
)
2026-02-14 12:15:19 +08:00