[Doc] Improve installation signposting (#12575)
- Make device tab names more explicit - Add comprehensive list of devices to https://docs.vllm.ai/en/latest/getting_started/installation/index.html - Add `attention` blocks to the intro of all devices that don't have pre-built wheels/images --------- Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
@@ -5,7 +5,8 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:selected:
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -15,7 +16,7 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
:::{include} rocm.inc.md
|
||||
@@ -25,7 +26,7 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
:::{include} xpu.inc.md
|
||||
@@ -45,7 +46,7 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -55,7 +56,7 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
:::{include} rocm.inc.md
|
||||
@@ -65,7 +66,7 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
:::{include} xpu.inc.md
|
||||
@@ -87,7 +88,7 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -97,14 +98,14 @@ vLLM is a Python library that supports the following GPU variants. Select your G
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
There is no extra information on creating a new Python environment for this device.
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
There is no extra information on creating a new Python environment for this device.
|
||||
@@ -118,7 +119,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -128,7 +129,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
:::{include} rocm.inc.md
|
||||
@@ -138,7 +139,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
:::{include} xpu.inc.md
|
||||
@@ -157,7 +158,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -167,7 +168,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
:::{include} rocm.inc.md
|
||||
@@ -177,7 +178,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
:::{include} xpu.inc.md
|
||||
@@ -196,7 +197,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -206,7 +207,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
:::{include} rocm.inc.md
|
||||
@@ -216,7 +217,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
:::{include} xpu.inc.md
|
||||
@@ -233,7 +234,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -243,7 +244,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
:::{include} rocm.inc.md
|
||||
@@ -253,7 +254,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
:::{include} xpu.inc.md
|
||||
@@ -270,7 +271,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
:::::{tab-set}
|
||||
:sync-group: device
|
||||
|
||||
::::{tab-item} CUDA
|
||||
::::{tab-item} NVIDIA CUDA
|
||||
:sync: cuda
|
||||
|
||||
:::{include} cuda.inc.md
|
||||
@@ -279,7 +280,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} ROCm
|
||||
::::{tab-item} AMD ROCm
|
||||
:sync: rocm
|
||||
|
||||
:::{include} rocm.inc.md
|
||||
@@ -288,7 +289,7 @@ There is no extra information on creating a new Python environment for this devi
|
||||
|
||||
::::
|
||||
|
||||
::::{tab-item} XPU
|
||||
::::{tab-item} Intel XPU
|
||||
:sync: xpu
|
||||
|
||||
:::{include} xpu.inc.md
|
||||
|
||||
@@ -2,6 +2,10 @@
|
||||
|
||||
vLLM supports AMD GPUs with ROCm 6.2.
|
||||
|
||||
:::{attention}
|
||||
There are no pre-built wheels for this device, so you must either use the pre-built Docker image or build vLLM from source.
|
||||
:::
|
||||
|
||||
## Requirements
|
||||
|
||||
- GPU: MI200s (gfx90a), MI300 (gfx942), Radeon RX 7900 series (gfx1100)
|
||||
@@ -13,14 +17,6 @@ vLLM supports AMD GPUs with ROCm 6.2.
|
||||
|
||||
Currently, there are no pre-built ROCm wheels.
|
||||
|
||||
However, the [AMD Infinity hub for vLLM](https://hub.docker.com/r/rocm/vllm/tags) offers a prebuilt, optimized
|
||||
docker image designed for validating inference performance on the AMD Instinct™ MI300X accelerator.
|
||||
|
||||
:::{tip}
|
||||
Please check [LLM inference performance validation on AMD Instinct MI300X](https://rocm.docs.amd.com/en/latest/how-to/performance-validation/mi300x/vllm-benchmark.html)
|
||||
for instructions on how to use this prebuilt docker image.
|
||||
:::
|
||||
|
||||
### Build wheel from source
|
||||
|
||||
0. Install prerequisites (skip if you are already in an environment/docker with the following installed):
|
||||
@@ -112,7 +108,13 @@ for instructions on how to use this prebuilt docker image.
|
||||
|
||||
### Pre-built images
|
||||
|
||||
Currently, there are no pre-built ROCm images.
|
||||
The [AMD Infinity hub for vLLM](https://hub.docker.com/r/rocm/vllm/tags) offers a prebuilt, optimized
|
||||
docker image designed for validating inference performance on the AMD Instinct™ MI300X accelerator.
|
||||
|
||||
:::{tip}
|
||||
Please check [LLM inference performance validation on AMD Instinct MI300X](https://rocm.docs.amd.com/en/latest/how-to/performance-validation/mi300x/vllm-benchmark.html)
|
||||
for instructions on how to use this prebuilt docker image.
|
||||
:::
|
||||
|
||||
### Build image from source
|
||||
|
||||
|
||||
@@ -2,6 +2,10 @@
|
||||
|
||||
vLLM initially supports basic model inferencing and serving on Intel GPU platform.
|
||||
|
||||
:::{attention}
|
||||
There are no pre-built wheels or images for this device, so you must build vLLM from source.
|
||||
:::
|
||||
|
||||
## Requirements
|
||||
|
||||
- Supported Hardware: Intel Data Center GPU, Intel ARC GPU
|
||||
|
||||
Reference in New Issue
Block a user