[Docs] Use 1-2-3 list for deploy steps in deployment/frameworks/ (#24633)

Signed-off-by: windsonsea <haifeng.yao@daocloud.io>
This commit is contained in:
Michael Yao
2025-09-11 16:50:12 +08:00
committed by GitHub
parent ba6011027d
commit d14c4ebf08
6 changed files with 98 additions and 90 deletions

View File

@@ -13,7 +13,7 @@ And LiteLLM supports all models on VLLM.
## Prerequisites
- Setup vLLM and litellm environment
Set up the vLLM and litellm environment:
```bash
pip install vllm litellm
@@ -23,13 +23,13 @@ pip install vllm litellm
### Chat completion
- Start the vLLM server with the supported chat completion model, e.g.
1. Start the vLLM server with the supported chat completion model, e.g.
```bash
vllm serve qwen/Qwen1.5-0.5B-Chat
```
```bash
vllm serve qwen/Qwen1.5-0.5B-Chat
```
- Call it with litellm:
1. Call it with litellm:
??? code
@@ -51,13 +51,13 @@ vllm serve qwen/Qwen1.5-0.5B-Chat
### Embeddings
- Start the vLLM server with the supported embedding model, e.g.
1. Start the vLLM server with the supported embedding model, e.g.
```bash
vllm serve BAAI/bge-base-en-v1.5
```
```bash
vllm serve BAAI/bge-base-en-v1.5
```
- Call it with litellm:
1. Call it with litellm:
```python
from litellm import embedding