You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/zero/gemini/__init__.py

21 lines
622 B

from .chunk import ChunkManager, TensorInfo, TensorState, search_chunk_configuration
from .colo_init_context import ColoInitContext, post_process_colo_init_ctx
from .gemini_ddp import GeminiDDP
from .gemini_mgr import GeminiManager
from .gemini_optimizer import GeminiAdamOptimizer, GeminiOptimizer
from .utils import get_static_torch_model
__all__ = [
"GeminiManager",
"TensorInfo",
"TensorState",
"ChunkManager",
"search_chunk_configuration",
"GeminiDDP",
"get_static_torch_model",
"GeminiAdamOptimizer",
"GeminiOptimizer",
"ColoInitContext",
"post_process_colo_init_ctx",
]