[Bugfix] support tie_word_embeddings for all models (#5724)
This commit is contained in:
@@ -260,6 +260,8 @@ class PhiForCausalLM(nn.Module, SupportsLoRA):
|
||||
super().__init__()
|
||||
|
||||
self.config = config
|
||||
# lm_head use bias, cannot share word embeddings
|
||||
assert not config.tie_word_embeddings
|
||||
self.lora_config = lora_config
|
||||
|
||||
self.quant_config = quant_config
|
||||
|
||||
Reference in New Issue
Block a user