[Minor] Fix small typo in llama.py: QKVParallelLinear -> QuantizationConfig (#4991)
This commit is contained in:
@@ -57,7 +57,7 @@ class LlamaMLP(nn.Module):
|
||||
hidden_size: int,
|
||||
intermediate_size: int,
|
||||
hidden_act: str,
|
||||
quant_config: Optional[QKVParallelLinear] = None,
|
||||
quant_config: Optional[QuantizationConfig] = None,
|
||||
bias: bool = False,
|
||||
) -> None:
|
||||
super().__init__()
|
||||
|
||||
Reference in New Issue
Block a user