Files
grace-gpu-containers/README.md

15 lines
674 B
Markdown
Raw Normal View History

2025-09-24 01:43:49 -05:00
# Building containers for GH200
Currently, prebuilt wheels for `vLLM` and `LMcache` are not available for `aarch64`. This can make setup tedious when working on modern `aarch64` platforms such as NVIDIA GH200.
This repository provides a Dockerfile to build a container with vLLM and all its dependencies pre-installed to try out various things such as KV offloading.
If you prefer not to build the image yourself, you can pull the ready-to-use image directly from Docker Hub:
2025-10-16 01:08:21 +00:00
```bash
docker run --rm -it --gpus all -v "$PWD":"$PWD" -w "$PWD" rajesh550/gh200-vllm:0.11.0 bash
```
2025-09-24 01:43:49 -05:00
👉 [Docker Hub](https://hub.docker.com/repository/docker/rajesh550/gh200-vllm/general)