2025-07-08 11:27:40 +01:00
|
|
|
# KServe
|
2024-12-23 17:35:38 -05:00
|
|
|
|
|
|
|
|
vLLM can be deployed with [KServe](https://github.com/kserve/kserve) on Kubernetes for highly scalable distributed model serving.
|
|
|
|
|
|
2025-11-24 09:40:05 -05:00
|
|
|
Please see [this guide](https://kserve.github.io/website/docs/model-serving/generative-inference/overview) for more details on using vLLM with KServe.
|