From 80d93fd6daf60d497c55a09c6dcd5471081c5978 Mon Sep 17 00:00:00 2001 From: pschlan-amd Date: Tue, 24 Feb 2026 04:08:34 +0100 Subject: [PATCH] gpu_model_runner: Cache is_encoder_decoder from model config (#35099) Signed-off-by: Patrick Schlangen --- vllm/config/model.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/config/model.py b/vllm/config/model.py index d7ff55205..5fb81ee42 100644 --- a/vllm/config/model.py +++ b/vllm/config/model.py @@ -1365,7 +1365,7 @@ class ModelConfig: return diff_sampling_param - @property + @cached_property def is_encoder_decoder(self) -> bool: """Extract the HF encoder/decoder model flag.""" return is_encoder_decoder(self.hf_config)