Fix CI failure - Flashinfer Kernel tests (#34316)
Signed-off-by: wzhao18 <wzhao18.sz@gmail.com>
This commit is contained in:
@@ -287,6 +287,7 @@ def test_flashinfer_cutlass_moe_fp8_no_graph(
|
|||||||
hidden_dim=k,
|
hidden_dim=k,
|
||||||
intermediate_size_per_partition=n,
|
intermediate_size_per_partition=n,
|
||||||
num_local_experts=e,
|
num_local_experts=e,
|
||||||
|
num_logical_experts=e,
|
||||||
activation=activation,
|
activation=activation,
|
||||||
device="cuda",
|
device="cuda",
|
||||||
moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(),
|
moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(),
|
||||||
|
|||||||
@@ -97,6 +97,7 @@ def test_flashinfer_fp4_moe_no_graph(
|
|||||||
hidden_dim=k,
|
hidden_dim=k,
|
||||||
intermediate_size_per_partition=n,
|
intermediate_size_per_partition=n,
|
||||||
num_local_experts=e,
|
num_local_experts=e,
|
||||||
|
num_logical_experts=e,
|
||||||
activation=activation,
|
activation=activation,
|
||||||
device="cuda",
|
device="cuda",
|
||||||
moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(),
|
moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(),
|
||||||
|
|||||||
@@ -147,6 +147,7 @@ def pplx_cutlass_moe(
|
|||||||
hidden_dim=hidden_dim,
|
hidden_dim=hidden_dim,
|
||||||
intermediate_size_per_partition=intermediate_dim,
|
intermediate_size_per_partition=intermediate_dim,
|
||||||
num_local_experts=num_local_experts,
|
num_local_experts=num_local_experts,
|
||||||
|
num_logical_experts=num_experts,
|
||||||
moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(),
|
moe_parallel_config=FusedMoEParallelConfig.make_no_parallel(),
|
||||||
activation="silu",
|
activation="silu",
|
||||||
in_dtype=torch.bfloat16,
|
in_dtype=torch.bfloat16,
|
||||||
|
|||||||
Reference in New Issue
Block a user