You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/legacy/nn/layer/parallel_2d/_utils.py

23 lines
853 B

from colossalai.legacy.context.parallel_mode import ParallelMode
from colossalai.legacy.core import global_context as gpc
from colossalai.legacy.global_variables import tensor_parallel_env as env
def get_summa_dim_from_env() -> int:
try:
summa_dim = env.summa_dim
assert summa_dim > 0, "SUMMA_DIM must be larger than zero"
return summa_dim
except KeyError:
raise EnvironmentError(
"SUMMA_DIM is not found in the current environment, "
"please make sure that you have used the correct process group initializer"
)
def assert_summa_initialization():
assert gpc.is_initialized(ParallelMode.PARALLEL_2D_COL) and gpc.is_initialized(
ParallelMode.PARALLEL_2D_ROW
), "Both TWO_DIMENSION_COL and TWO_DIMENSION_ROW must be initialized by the process group initializer"