You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/legacy/nn/layer/parallel_2p5d/__init__.py

23 lines
494 B

from ._operation import reduce_by_batch_2p5d, split_batch_2p5d
from .layers import (
Classifier2p5D,
Embedding2p5D,
LayerNorm2p5D,
Linear2p5D,
PatchEmbedding2p5D,
VocabParallelClassifier2p5D,
VocabParallelEmbedding2p5D,
)
__all__ = [
"split_batch_2p5d",
"reduce_by_batch_2p5d",
"Linear2p5D",
"LayerNorm2p5D",
"Classifier2p5D",
"PatchEmbedding2p5D",
"Embedding2p5D",
"VocabParallelClassifier2p5D",
"VocabParallelEmbedding2p5D",
]