[Kernel][Core] Add AWQ support to the Marlin kernel (#6612)
This commit is contained in:
committed by
GitHub
parent
25e778aa16
commit
396d92d5e0
@@ -251,7 +251,7 @@ class ModelConfig:
|
||||
f"supported in ROCm.")
|
||||
if (self.quantization
|
||||
not in ("fp8", "marlin", "gptq_marlin_24", "gptq_marlin",
|
||||
"fbgemm_fp8", "compressed_tensors")):
|
||||
"awq_marlin", "fbgemm_fp8", "compressed_tensors")):
|
||||
logger.warning(
|
||||
"%s quantization is not fully "
|
||||
"optimized yet. The speed can be slower than "
|
||||
|
||||
Reference in New Issue
Block a user