Stop using title frontmatter and fix doc that can only be reached by search (#20623)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
Harry Mellor
2025-07-08 11:27:40 +01:00
committed by GitHub
parent b4bab81660
commit b942c094e3
81 changed files with 82 additions and 238 deletions

View File

@@ -1,6 +1,4 @@
---
title: Offline Inference
---
# Offline Inference
Offline inference is possible in your own code using vLLM's [`LLM`][vllm.LLM] class.
@@ -23,7 +21,7 @@ The available APIs depend on the model type:
!!! info
[API Reference][offline-inference-api]
### Ray Data LLM API
## Ray Data LLM API
Ray Data LLM is an alternative offline inference API that uses vLLM as the underlying engine.
This API adds several batteries-included capabilities that simplify large-scale, GPU-efficient inference: