Fix Mistral config remap to accept compressed-tensors quantization #34028 (#34104)

Signed-off-by: baonudesifeizhai <baonudesifeizhai@gmail.com>
This commit is contained in:
baonudesifeizhai
2026-02-12 03:22:06 -05:00
committed by GitHub
parent 55a1a9563a
commit f5897613fb

View File

@@ -198,6 +198,14 @@ def _remap_mistral_quantization_args(config: dict) -> dict:
"quant_method": "fp8",
"activation_scheme": "dynamic" if is_dynamic else "static",
}
elif (
str(quantization.get("quant_method", "")).lower().replace("_", "-")
== "compressed-tensors"
):
# Pass through compressed-tensors config, while normalizing
# quant_method to the canonical community spelling.
quantization["quant_method"] = "compressed-tensors"
config["quantization_config"] = quantization
else:
raise ValueError(f"Found unknown quantization='{quantization}' in config")