[ROCm][Bugfix] Fix Aiter RMSNorm (#23412)
Signed-off-by: vllmellm <vllm.ellm@embeddedllm.com>
This commit is contained in:
@@ -9,11 +9,11 @@ import torch.nn as nn
|
||||
import vllm.envs as envs
|
||||
from vllm.model_executor.custom_op import CustomOp
|
||||
from vllm.platforms import current_platform
|
||||
from vllm.utils import direct_register_custom_op
|
||||
|
||||
|
||||
def is_rocm_aiter_rmsnorm_enabled() -> bool:
|
||||
return current_platform.is_rocm() \
|
||||
and envs.VLLM_ROCM_USE_AITER_RMSNORM \
|
||||
return envs.VLLM_ROCM_USE_AITER_RMSNORM \
|
||||
and envs.VLLM_ROCM_USE_AITER
|
||||
|
||||
|
||||
@@ -43,8 +43,8 @@ def fused_add_rms_norm(
|
||||
return x, residual
|
||||
|
||||
|
||||
def rocm_aiter_rms_norm(x: torch.Tensor, weight: torch.Tensor,
|
||||
variance_epsilon: float) -> torch.Tensor:
|
||||
def rocm_aiter_rms_norm_impl(x: torch.Tensor, weight: torch.Tensor,
|
||||
variance_epsilon: float) -> torch.Tensor:
|
||||
import aiter as rocm_aiter
|
||||
if x.dim() > 2:
|
||||
x_original_shape = x.shape
|
||||
@@ -55,7 +55,7 @@ def rocm_aiter_rms_norm(x: torch.Tensor, weight: torch.Tensor,
|
||||
return rocm_aiter.rms_norm(x, weight, variance_epsilon)
|
||||
|
||||
|
||||
def rocm_aiter_fused_add_rms_norm(
|
||||
def rocm_aiter_rmsnorm2d_fwd_with_add_impl(
|
||||
x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor,
|
||||
variance_epsilon: float) -> tuple[torch.Tensor, torch.Tensor]:
|
||||
|
||||
@@ -74,14 +74,48 @@ def rocm_aiter_fused_add_rms_norm(
|
||||
return output, residual_out
|
||||
|
||||
|
||||
def dispatch_cuda_rmsnorm_func(add_residual: bool):
|
||||
if add_residual:
|
||||
if is_rocm_aiter_rmsnorm_enabled():
|
||||
return rocm_aiter_fused_add_rms_norm
|
||||
return fused_add_rms_norm
|
||||
def rocm_aiter_rms_norm_fake(x: torch.Tensor, weight: torch.Tensor,
|
||||
variance_epsilon: float) -> torch.Tensor:
|
||||
return torch.empty_like(x)
|
||||
|
||||
if is_rocm_aiter_rmsnorm_enabled():
|
||||
return rocm_aiter_rms_norm
|
||||
|
||||
def rocm_aiter_rmsnorm2d_fwd_with_add_fake(
|
||||
x: torch.Tensor, residual: torch.Tensor, weight: torch.Tensor,
|
||||
variance_epsilon: float) -> tuple[torch.Tensor, torch.Tensor]:
|
||||
return torch.empty_like(x), torch.empty_like(residual)
|
||||
|
||||
|
||||
if current_platform.is_rocm():
|
||||
direct_register_custom_op(
|
||||
op_name="rocm_aiter_rms_norm",
|
||||
op_func=rocm_aiter_rms_norm_impl,
|
||||
mutates_args=[],
|
||||
fake_impl=rocm_aiter_rms_norm_fake,
|
||||
dispatch_key=current_platform.dispatch_key,
|
||||
)
|
||||
|
||||
direct_register_custom_op(
|
||||
op_name="rocm_aiter_rmsnorm2d_fwd_with_add",
|
||||
op_func=rocm_aiter_rmsnorm2d_fwd_with_add_impl,
|
||||
mutates_args=[],
|
||||
fake_impl=rocm_aiter_rmsnorm2d_fwd_with_add_fake,
|
||||
dispatch_key=current_platform.dispatch_key,
|
||||
)
|
||||
|
||||
|
||||
def dispatch_rocm_rmsnorm_func(with_fused_add: bool, dtype: torch.dtype):
|
||||
use_aiter = is_rocm_aiter_rmsnorm_enabled() and dtype in [
|
||||
torch.float16, torch.bfloat16
|
||||
]
|
||||
|
||||
if use_aiter and with_fused_add:
|
||||
return torch.ops.vllm.rocm_aiter_rmsnorm2d_fwd_with_add
|
||||
if use_aiter:
|
||||
return torch.ops.vllm.rocm_aiter_rms_norm
|
||||
|
||||
# fall back to CUDA implementation
|
||||
if with_fused_add:
|
||||
return fused_add_rms_norm
|
||||
return rms_norm
|
||||
|
||||
|
||||
@@ -114,6 +148,13 @@ class RMSNorm(CustomOp):
|
||||
self.weight = torch.ones(hidden_size)
|
||||
if self.has_weight:
|
||||
self.weight = nn.Parameter(self.weight)
|
||||
weight_dtype = self.weight.data.dtype
|
||||
|
||||
if current_platform.is_rocm():
|
||||
self.rocm_norm_func = dispatch_rocm_rmsnorm_func(
|
||||
with_fused_add=False, dtype=weight_dtype)
|
||||
self.rocm_norm_func_with_add = dispatch_rocm_rmsnorm_func(
|
||||
with_fused_add=True, dtype=weight_dtype)
|
||||
|
||||
def forward_native(
|
||||
self,
|
||||
@@ -162,13 +203,27 @@ class RMSNorm(CustomOp):
|
||||
return self.forward_native(x, residual)
|
||||
|
||||
add_residual = residual is not None
|
||||
norm_func = dispatch_cuda_rmsnorm_func(add_residual)
|
||||
|
||||
if add_residual:
|
||||
return norm_func(x, residual, self.weight.data,
|
||||
self.variance_epsilon)
|
||||
return fused_add_rms_norm(x, residual, self.weight.data,
|
||||
self.variance_epsilon)
|
||||
else:
|
||||
return norm_func(x, self.weight.data, self.variance_epsilon)
|
||||
return rms_norm(x, self.weight.data, self.variance_epsilon)
|
||||
|
||||
def forward_hip(
|
||||
self,
|
||||
x: torch.Tensor,
|
||||
residual: Optional[torch.Tensor] = None,
|
||||
) -> Union[torch.Tensor, tuple[torch.Tensor, torch.Tensor]]:
|
||||
if self.variance_size_override is not None:
|
||||
return self.forward_native(x, residual)
|
||||
|
||||
add_residual = residual is not None
|
||||
if add_residual:
|
||||
return self.rocm_norm_func_with_add(x, residual, self.weight.data,
|
||||
self.variance_epsilon)
|
||||
else:
|
||||
return self.rocm_norm_func(x, self.weight.data,
|
||||
self.variance_epsilon)
|
||||
|
||||
def forward_xpu(
|
||||
self,
|
||||
|
||||
Reference in New Issue
Block a user