mirror of https://github.com/InternLM/InternLM
fix logger for moe
parent
36d1bd2f41
commit
3df0a51555
|
@ -365,11 +365,12 @@ def launch(
|
||||||
f"data parallel size: {gpc.data_parallel_size}, pipeline parallel size: {gpc.pipeline_parallel_size}, "
|
f"data parallel size: {gpc.data_parallel_size}, pipeline parallel size: {gpc.pipeline_parallel_size}, "
|
||||||
f"tensor parallel size: {gpc.tensor_parallel_size}",
|
f"tensor parallel size: {gpc.tensor_parallel_size}",
|
||||||
)
|
)
|
||||||
logger.info(
|
if hasattr(gpc.config.model, "num_experts") and gpc.config.model.num_experts > 1:
|
||||||
f"Creating MoE with num_experts: {gpc.config.model.num_experts} | "
|
logger.info(
|
||||||
f"expert parallel size: {gpc.expert_parallel_size} | "
|
f"Creating MoE with num_experts: {gpc.config.model.num_experts} | "
|
||||||
f"number of local experts: {gpc.config.model.num_experts//gpc.expert_parallel_size}"
|
f"expert parallel size: {gpc.expert_parallel_size} | "
|
||||||
)
|
f"number of local experts: {gpc.config.model.num_experts//gpc.expert_parallel_size}"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def launch_from_slurm(
|
def launch_from_slurm(
|
||||||
|
|
Loading…
Reference in New Issue