Fix spelling errors (#33978)
This commit is contained in:
@@ -139,7 +139,7 @@ def prepare_inputs(
|
||||
masked_m.append(mask.sum())
|
||||
|
||||
masked_m = torch.tensor(masked_m, dtype=torch.int32)
|
||||
# Intialize the hidden_states_3d with ones instead of empty to avoid nan
|
||||
# Initialize the hidden_states_3d with ones instead of empty to avoid nan
|
||||
# issue.
|
||||
hidden_states_3d = torch.ones(
|
||||
(num_experts, max(masked_m), hidden_states.shape[1]), dtype=hidden_states.dtype
|
||||
|
||||
@@ -334,7 +334,7 @@ def test_batched_moe_align_block_size(
|
||||
ref_expert_ids = torch.empty((Msum // block_size,), dtype=torch.int32)
|
||||
ref_num_tokens_post_pad = torch.empty((1,), dtype=torch.int32)
|
||||
|
||||
# Intialize
|
||||
# Initialize
|
||||
sentinel = E * max_tokens_per_batch
|
||||
ref_sorted_ids.fill_(sentinel)
|
||||
ref_expert_ids.fill_(-1)
|
||||
|
||||
Reference in New Issue
Block a user