From fc56f4a0714bd1727ea100a9d75d18ce1a138684 Mon Sep 17 00:00:00 2001 From: Fadi Arafeh <115173828+fadara01@users.noreply.github.com> Date: Thu, 22 Jan 2026 22:27:40 +0000 Subject: [PATCH] [BugFix] Fix invalid flashinfer_fused_moe_blockscale_fp8 op registration (#32855) Signed-off-by: Fadi Arafeh --- vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py b/vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py index ad9eb0d40..647108cc4 100644 --- a/vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py +++ b/vllm/model_executor/layers/fused_moe/flashinfer_trtllm_moe.py @@ -131,7 +131,7 @@ def flashinfer_fused_moe_blockscale_fp8( expert_offset: int, local_num_experts: int, block_shape: list[int], - routing_method_type: int = RoutingMethodType.DeepSeekV3, + routing_method_type: int = int(RoutingMethodType.DeepSeekV3), routed_scaling: float | None = 1.0, ) -> torch.Tensor: from vllm.utils.flashinfer import flashinfer_trtllm_fp8_block_scale_moe