From 3a7cd627a86c61a6dbc8d365c2e3c921b94e9971 Mon Sep 17 00:00:00 2001 From: Siyuan Liu Date: Mon, 9 Jun 2025 16:41:51 -0700 Subject: [PATCH] [Misc] Fix a config typo in disable_hybrid_kv_cache_manager configuration (#19383) Signed-off-by: Siyuan Liu --- vllm/config.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/vllm/config.py b/vllm/config.py index d8318ab22..3fbb6015f 100644 --- a/vllm/config.py +++ b/vllm/config.py @@ -4497,13 +4497,13 @@ class VllmConfig: # warning message here and will log it later. if not (current_platform.is_cuda() or current_platform.is_rocm()): # Hybrid KV cache manager is not supported on non-GPU platforms. - self.disable_hybrid_kv_cache_manager = True + self.scheduler_config.disable_hybrid_kv_cache_manager = True if self.kv_transfer_config is not None: # Hybrid KV cache manager is not compatible with KV transfer. - self.disable_hybrid_kv_cache_manager = True + self.scheduler_config.disable_hybrid_kv_cache_manager = True if self.kv_events_config is not None: # Hybrid KV cache manager is not compatible with KV events. - self.disable_hybrid_kv_cache_manager = True + self.scheduler_config.disable_hybrid_kv_cache_manager = True def update_sizes_for_sequence_parallelism(self, possible_sizes: list) -> list: