[Bugfix] Fix Positive Feature Layers in Llava Models (#13514)
Signed-off-by: Alex-Brooks <Alex.brooks@ibm.com>
This commit is contained in:
@@ -969,7 +969,7 @@ class PixtralHFTransformer(nn.Module):
|
||||
position_embeddings: torch.Tensor,
|
||||
return_all_hidden_states: bool,
|
||||
) -> torch.Tensor:
|
||||
hidden_states_pool = []
|
||||
hidden_states_pool = [x]
|
||||
|
||||
for layer in self.layers:
|
||||
x = layer(x, attention_mask, position_embeddings)
|
||||
|
||||
Reference in New Issue
Block a user