mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
23 lines
853 B
23 lines
853 B
from colossalai.legacy.context.parallel_mode import ParallelMode
|
|
from colossalai.legacy.core import global_context as gpc
|
|
from colossalai.legacy.global_variables import tensor_parallel_env as env
|
|
|
|
|
|
def get_summa_dim_from_env() -> int:
|
|
try:
|
|
summa_dim = env.summa_dim
|
|
assert summa_dim > 0, "SUMMA_DIM must be larger than zero"
|
|
return summa_dim
|
|
|
|
except KeyError:
|
|
raise EnvironmentError(
|
|
"SUMMA_DIM is not found in the current environment, "
|
|
"please make sure that you have used the correct process group initializer"
|
|
)
|
|
|
|
|
|
def assert_summa_initialization():
|
|
assert gpc.is_initialized(ParallelMode.PARALLEL_2D_COL) and gpc.is_initialized(
|
|
ParallelMode.PARALLEL_2D_ROW
|
|
), "Both TWO_DIMENSION_COL and TWO_DIMENSION_ROW must be initialized by the process group initializer"
|