[Model] Switch to Fused RMS norm in Qwen2.5_VL model. (#22184)
Signed-off-by: kf <kuanfu.liu@embeddedllm.com> Signed-off-by: tjtanaavllm <tunjian.tan@amd.com> Signed-off-by: vllmellm <vllm.ellm@embeddedllm.com> Co-authored-by: kf <kuanfu.liu@embeddedllm.com>
This commit is contained in:
@@ -396,13 +396,13 @@ class Qwen2_5_VisionBlock(nn.Module):
|
|||||||
max_seqlen: Optional[int] = None, # Only used for Flash Attention
|
max_seqlen: Optional[int] = None, # Only used for Flash Attention
|
||||||
seqlens: Optional[list[int]] = None, # Only used for xFormers
|
seqlens: Optional[list[int]] = None, # Only used for xFormers
|
||||||
) -> torch.Tensor:
|
) -> torch.Tensor:
|
||||||
x = x + self.attn(self.norm1(x),
|
x_attn = self.attn(self.norm1(x),
|
||||||
cu_seqlens=cu_seqlens,
|
cu_seqlens=cu_seqlens,
|
||||||
rotary_pos_emb=rotary_pos_emb,
|
rotary_pos_emb=rotary_pos_emb,
|
||||||
max_seqlen=max_seqlen,
|
max_seqlen=max_seqlen,
|
||||||
seqlens=seqlens)
|
seqlens=seqlens)
|
||||||
|
x_fused_norm, residual = self.norm2(x, residual=x_attn)
|
||||||
x = x + self.mlp(self.norm2(x))
|
x = residual + self.mlp(x_fused_norm)
|
||||||
return x
|
return x
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user