[cherry-pick][Bugfix] Fix EP weight filter breaking EPLB and NVFP4 accuracy #37322

Signed-off-by: khluu <khluu000@gmail.com>
This commit is contained in:
khluu
2026-03-18 01:48:32 -07:00
parent e60c1674b3
commit 262ddd0d81
2 changed files with 12 additions and 0 deletions

View File

@@ -319,6 +319,13 @@ class DefaultModelLoader(BaseModelLoader):
and parallel_config.enable_ep_weight_filter
):
return
# When EPLB is enabled, redundant physical expert slots may map to
# logical experts that belong to other ranks in the default partition.
# The weight loader needs to see ALL logical expert weights so it can
# populate these redundant slots. Skip the filter entirely.
if parallel_config.enable_eplb:
return
num_experts = model_config.get_num_experts()
if num_experts <= 0:

View File

@@ -73,4 +73,9 @@ def should_skip_weight(
if eid is None:
# Not an expert weight (dense / shared-expert / embedding) → keep.
return False
# Only skip heavy weight tensors, never scale/metadata tensors.
# Scale tensors are tiny and some backends need them from ALL experts
# (e.g. FlashInfer NVFP4 computes a global max of activation scales).
if not weight_name.endswith(".weight"):
return False
return eid not in local_expert_ids