[Quantization] Enable compressed-tensors AWQ for Turing GPU (#29732)
Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
This commit is contained in:
@@ -79,8 +79,8 @@ class CompressedTensorsWNA16(CompressedTensorsScheme):
|
||||
|
||||
@classmethod
|
||||
def get_min_capability(cls) -> int:
|
||||
# ampere and up
|
||||
return 80
|
||||
# Turing and up
|
||||
return 75
|
||||
|
||||
def create_weights(
|
||||
self,
|
||||
|
||||
Reference in New Issue
Block a user