From 014faf6c5a072fd27d9df945a02f90755a369ecc Mon Sep 17 00:00:00 2001 From: botbw Date: Fri, 12 Jul 2024 03:29:16 +0000 Subject: [PATCH] [chore] manually revert unintended commit --- colossalai/booster/plugin/moe_hybrid_parallel_plugin.py | 8 -------- 1 file changed, 8 deletions(-) diff --git a/colossalai/booster/plugin/moe_hybrid_parallel_plugin.py b/colossalai/booster/plugin/moe_hybrid_parallel_plugin.py index 902500e42..22157b5cf 100644 --- a/colossalai/booster/plugin/moe_hybrid_parallel_plugin.py +++ b/colossalai/booster/plugin/moe_hybrid_parallel_plugin.py @@ -65,14 +65,6 @@ class MoeHybridParallelZeroOptimizer(HybridParallelZeroOptimizer): overlap_communication = True warnings.warn(WARN_STR + " Please make sure of this.") - self.param_info = param_info - self.stage_manager = model.stage_manager - self.shared_params = model.shared_params - self.dp_pg = dp_process_group - - if use_pipeline: - reinitialize_optimizer(optimizer, model) - pg_param_list = { dp_process_group: list(filter(lambda p: not is_moe_tensor(p), model.parameters())), moe_dp_group: list(filter(is_moe_tensor, model.parameters())),