Fix eplb nvfp4 experts hook (#37217)

Signed-off-by: Elvir Crncevic <elvircrn@gmail.com>
Signed-off-by: Elvir Crncevic <elvir@anthropic.com>
Co-authored-by: Tyler Michael Smith <tyler@neuralmagic.com>
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
Elvir Crnčević
2026-03-16 23:03:54 +01:00
committed by GitHub
parent c0f011918d
commit fd4d96302a
10 changed files with 57 additions and 25 deletions

View File

@@ -489,6 +489,9 @@ class FusedMoEExperts(ABC):
self.max_num_tokens = max_num_tokens
self.num_dispatchers = num_dispatchers
def process_weights_after_loading(self, layer: torch.nn.Module) -> None: # noqa: B027
pass
@staticmethod
def is_monolithic() -> bool:
raise NotImplementedError("Implemented by subclasses.")