This website requires JavaScript.
Explore
Help
Register
Sign In
biondizzle
/
vllm
Watch
1
Star
0
Fork
0
You've already forked vllm
Code
Issues
Pull Requests
Actions
2
Packages
Projects
Releases
Wiki
Activity
Files
e23f665d835a40cfe2fbf21df76a4cbdd3d60b4c
vllm
/
docs
/
serving
History
Ben Browning
e1dd706cd1
[Frontend] Respect Chat Completion parallel_tool_calls param (
#26233
)
...
Signed-off-by: Ben Browning <
bbrownin@redhat.com
> Co-authored-by: Chauncey <
chaunceyjiang@gmail.com
>
2025-11-25 09:56:15 +00:00
..
integrations
[Doc] ruff format remaining Python examples (
#26795
)
2025-10-15 01:25:49 -07:00
context_parallel_deployment.md
[doc] add Context Parallel Deployment doc (
#26877
)
2025-10-15 16:33:52 +08:00
data_parallel_deployment.md
[Data-parallel] Allow DP>1 for world_size > num_gpus on node (8) (
#26367
)
2025-10-17 08:24:42 -07:00
distributed_troubleshooting.md
[Docs] Replace all explicit anchors with real links (
#27087
)
2025-10-17 02:22:06 -07:00
expert_parallel_deployment.md
[Docs] Reduce custom syntax used in docs (
#27009
)
2025-10-16 20:05:34 -07:00
offline_inference.md
[Docs] Replace all explicit anchors with real links (
#27087
)
2025-10-17 02:22:06 -07:00
openai_compatible_server.md
[Frontend] Respect Chat Completion parallel_tool_calls param (
#26233
)
2025-11-25 09:56:15 +00:00
parallelism_scaling.md
docs: fixes distributed executor backend config for multi-node vllm (
#29173
)
2025-11-23 10:58:28 +08:00