diff --git a/vllm/model_executor/models/glm4_1v.py b/vllm/model_executor/models/glm4_1v.py index ff76a26bb..4722b6e3d 100644 --- a/vllm/model_executor/models/glm4_1v.py +++ b/vllm/model_executor/models/glm4_1v.py @@ -63,6 +63,9 @@ from vllm.model_executor.layers.linear import ( RowParallelLinear, ) from vllm.model_executor.layers.quantization import QuantizationConfig +from vllm.model_executor.layers.quantization.compressed_tensors import ( + compressed_tensors, +) from vllm.model_executor.layers.rotary_embedding import get_rope from vllm.model_executor.layers.rotary_embedding.common import ( ApplyRotaryEmb, @@ -280,7 +283,9 @@ class Glm4vVisionAttention(nn.Module): bias=False, quant_config=quant_config, # Change qkv prefix to align with GLM-4.5V-FP8 quantization cfg - prefix=f"{prefix}.qkv_proj" if quant_config else f"{prefix}.qkv", + prefix=f"{prefix}.qkv_proj" + if isinstance(quant_config, compressed_tensors.CompressedTensorsConfig) + else f"{prefix}.qkv", disable_tp=use_data_parallel, ) self.proj = RowParallelLinear(