[Docs] Improve API docs (+small tweaks) (#22459)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
Harry Mellor
2025-08-08 11:02:51 +01:00
committed by GitHub
parent ccdae737a0
commit 7be7f3824a
5 changed files with 5 additions and 13 deletions

View File

@@ -1,7 +1,4 @@
---
title: FP8 INC
---
[](){ #inc }
# FP8 INC
vLLM supports FP8 (8-bit floating point) weight and activation quantization using Intel® Neural Compressor (INC) on Intel® Gaudi® 2 and Intel® Gaudi® 3 AI accelerators.
Currently, quantization is validated only in Llama models.