[Bugfix] Fix embedding to support 2D inputs (#5829)
This commit is contained in:
@@ -310,7 +310,7 @@ class VocabParallelEmbedding(torch.nn.Module):
|
|||||||
output_parallel = F.embedding(masked_input.long(), self.weight)
|
output_parallel = F.embedding(masked_input.long(), self.weight)
|
||||||
# Mask the output embedding.
|
# Mask the output embedding.
|
||||||
if self.tp_size > 1:
|
if self.tp_size > 1:
|
||||||
output_parallel.masked_fill_(input_mask.unsqueeze(1), 0)
|
output_parallel.masked_fill_(input_mask.unsqueeze(-1), 0)
|
||||||
# Reduce across all the model parallel GPUs.
|
# Reduce across all the model parallel GPUs.
|
||||||
output = tensor_model_parallel_all_reduce(output_parallel)
|
output = tensor_model_parallel_all_reduce(output_parallel)
|
||||||
return output
|
return output
|
||||||
|
|||||||
Reference in New Issue
Block a user