[Doc] Add engine args back in to the docs (#20674)

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
Harry Mellor
2025-07-10 16:02:40 +01:00
committed by GitHub
parent 77f77a951e
commit 3482fd7e4e
14 changed files with 218 additions and 40 deletions

View File

@@ -16,15 +16,18 @@ from transformers import (AutoTokenizer, PreTrainedTokenizer,
from vllm import envs
from vllm.logger import init_logger
from vllm.lora.request import LoRARequest
from vllm.transformers_utils.tokenizer_base import (TokenizerBase,
TokenizerRegistry)
from vllm.transformers_utils.tokenizers import MistralTokenizer
from vllm.transformers_utils.utils import check_gguf_file
from vllm.utils import make_async
if TYPE_CHECKING:
from vllm.config import ModelConfig
from vllm.lora.request import LoRARequest
from vllm.transformers_utils.tokenizer_base import TokenizerBase
else:
ModelConfig = Any
LoRARequest = Any
TokenizerBase = Any
logger = init_logger(__name__)
@@ -222,6 +225,7 @@ def get_tokenizer(
tokenizer = MistralTokenizer.from_pretrained(str(tokenizer_name),
revision=revision)
elif tokenizer_mode == "custom":
from vllm.transformers_utils.tokenizer_base import TokenizerRegistry
tokenizer = TokenizerRegistry.get_tokenizer(str(tokenizer_name),
*args,
revision=revision,
@@ -271,7 +275,7 @@ cached_get_tokenizer = lru_cache(get_tokenizer)
def cached_tokenizer_from_config(
model_config: "ModelConfig",
model_config: ModelConfig,
**kwargs: Any,
):
return cached_get_tokenizer(