[Minor] Rename quantization nvfp4 to modelopt_fp4 (#18356)

Signed-off-by: mgoin <mgoin64@gmail.com>
This commit is contained in:
Michael Goin
2025-05-20 12:08:37 -04:00
committed by GitHub
parent 8f55962a7f
commit f4a8a37465
4 changed files with 7 additions and 7 deletions

View File

@@ -824,7 +824,7 @@ class ModelConfig:
optimized_quantization_methods = [
"fp8", "marlin", "modelopt", "gptq_marlin_24", "gptq_marlin",
"awq_marlin", "fbgemm_fp8", "compressed-tensors", "experts_int8",
"quark", "nvfp4", "bitblas", "gptq_bitblas"
"quark", "modelopt_fp4", "bitblas", "gptq_bitblas"
]
if self.quantization is not None:
self.quantization = cast(QuantizationMethods,