[Feature] use --eplb_config to set eplb param (#20562)
Signed-off-by: rongfu.leng <rongfu.leng@daocloud.io> Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com> Signed-off-by: rongfu.leng <lenronfu@gmail.com> Co-authored-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
This commit is contained in:
@@ -515,7 +515,7 @@ class Worker(WorkerBase):
|
||||
assert self.model_runner.eplb_state is not None
|
||||
new_physical_experts = \
|
||||
self.model_runner.eplb_state.physical_to_logical_map.shape[1]
|
||||
parallel_config.num_redundant_experts = (
|
||||
parallel_config.eplb_config.num_redundant_experts = (
|
||||
new_physical_experts -
|
||||
self.model_runner.eplb_state.logical_replica_count.shape[1])
|
||||
global_expert_load = None
|
||||
@@ -531,7 +531,7 @@ class Worker(WorkerBase):
|
||||
assert self.model_runner.eplb_state is not None
|
||||
global_expert_load = self.model_runner.eplb_state.rearrange(
|
||||
self.model_runner.model, execute_shuffle=False)
|
||||
parallel_config.num_redundant_experts = (
|
||||
parallel_config.eplb_config.num_redundant_experts = (
|
||||
new_physical_experts - global_expert_load.shape[1])
|
||||
prepare_communication_buffer_for_model(self.model_runner.model)
|
||||
self.model_runner.model.update_physical_experts_metadata(
|
||||
|
||||
Reference in New Issue
Block a user