You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/legacy/context/process_group_initializer/__init__.py

24 lines
763 B

from .initializer_1d import Initializer_1D
from .initializer_2d import Initializer_2D
from .initializer_2p5d import Initializer_2p5D
from .initializer_3d import Initializer_3D
from .initializer_data import Initializer_Data
from .initializer_model import Initializer_Model
from .initializer_pipeline import Initializer_Pipeline
from .initializer_sequence import Initializer_Sequence
from .initializer_tensor import Initializer_Tensor
from .process_group_initializer import ProcessGroupInitializer
__all__ = [
"Initializer_Tensor",
"Initializer_Sequence",
"Initializer_Pipeline",
"Initializer_Data",
"Initializer_2p5D",
"Initializer_2D",
"Initializer_3D",
"Initializer_1D",
"ProcessGroupInitializer",
"Initializer_Model",
]