diff --git a/vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors.py b/vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors.py index 2e61b0609..330d79110 100644 --- a/vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors.py +++ b/vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors.py @@ -751,11 +751,7 @@ class CompressedTensorsConfig(QuantizationConfig): model_compression_config=model_compression_config, ) elif weight_quant is None: - logger.warning_once( - "Acceleration for non-quantized schemes is " - "not supported by Compressed Tensors. " - "Falling back to UnquantizedLinearMethod" - ) + # Falling back to UnquantizedLinearMethod return None else: