[Misc] Remove redundant line (#32366)
Signed-off-by: wangli <wangli858794774@gmail.com>
This commit is contained in:
@@ -363,7 +363,6 @@ class Attention(nn.Module, AttentionLayerBase):
|
||||
output_shape = torch.Size(
|
||||
(num_tokens, self.num_heads * self.head_size_v)
|
||||
)
|
||||
output_shape = output_shape if output_shape is not None else query.shape
|
||||
output = torch.empty(output_shape, dtype=output_dtype, device=query.device)
|
||||
hidden_size = output_shape[-1]
|
||||
# Reshape the query, key, and value tensors.
|
||||
|
||||
Reference in New Issue
Block a user