[Misc] Remove redundant line (#32366)

Signed-off-by: wangli <wangli858794774@gmail.com>
This commit is contained in:
Li Wang
2026-01-15 12:29:56 +08:00
committed by GitHub
parent 375e5984fe
commit d86fc23bdd

View File

@@ -363,7 +363,6 @@ class Attention(nn.Module, AttentionLayerBase):
output_shape = torch.Size(
(num_tokens, self.num_heads * self.head_size_v)
)
output_shape = output_shape if output_shape is not None else query.shape
output = torch.empty(output_shape, dtype=output_dtype, device=query.device)
hidden_size = output_shape[-1]
# Reshape the query, key, and value tensors.