[Minor] Fix small typo in llama.py: QKVParallelLinear -> QuantizationConfig (#4991)

This commit is contained in:
Philipp Moritz
2024-05-22 15:26:56 -07:00
committed by GitHub
parent eb6d3c264d
commit a36de682d4

View File

@@ -57,7 +57,7 @@ class LlamaMLP(nn.Module):
hidden_size: int,
intermediate_size: int,
hidden_act: str,
quant_config: Optional[QKVParallelLinear] = None,
quant_config: Optional[QuantizationConfig] = None,
bias: bool = False,
) -> None:
super().__init__()