biondizzle 10c71a446c Remove flash-attn GIT_TAG override to main — causes FLASHATTENTION_FP8_TWO_LEVEL_INTERVAL undefined error
v0.19.0 pins a compatible flash-attn commit (2921022). The sed that
forced GIT_TAG to main pulled in newer code that references
FLASHATTENTION_FP8_TWO_LEVEL_INTERVAL which isn't defined in v0.19.0's
build config. Use the pinned commit instead.
2026-04-28 03:07:14 +00:00
2026-04-03 04:27:21 +00:00
2025-10-23 18:16:57 +00:00

Building containers for GH200

Currently, prebuilt wheels for vLLM and LMcache are not available for aarch64. This can make setup tedious when working on modern aarch64 platforms such as NVIDIA GH200.

Further, Nvidia at this time does not provide the Dockerfile associated with the NGC containers which makes replacing some of the components (like a newer version of vLLM) tedious.

This repository provides a Dockerfile to build a container with vLLM and all its dependencies pre-installed to try out various things such as KV offloading.

If you prefer not to build the image yourself, you can pull the ready-to-use image directly from Docker Hub:

docker run --rm -it --gpus all -v "$PWD":"$PWD" -w "$PWD" rajesh550/gh200-vllm:0.11.0 bash

# CUDA 13
docker run --rm -it --gpus all -v "$PWD":"$PWD" -w "$PWD" rajesh550/gh200-vllm:0.11.1rc2 bash

👉 Docker Hub

Version info:

CUDA: 13.0.1
Ubuntu: 24.04
Python: 3.12
PyTorch: 2.9.0+cu130
Triton: 3.5.x
xformers: 0.32.post2+
flashinfer: 0.4.1
flashattention: 3.0.0b1
LMCache: 0.3.7
vLLM: 0.11.1rc3
Description
No description provided
Readme 6.4 MiB
Languages
Dockerfile 51.3%
Python 48.7%