mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
23 lines
494 B
23 lines
494 B
from ._operation import reduce_by_batch_2p5d, split_batch_2p5d
|
|
from .layers import (
|
|
Classifier2p5D,
|
|
Embedding2p5D,
|
|
LayerNorm2p5D,
|
|
Linear2p5D,
|
|
PatchEmbedding2p5D,
|
|
VocabParallelClassifier2p5D,
|
|
VocabParallelEmbedding2p5D,
|
|
)
|
|
|
|
__all__ = [
|
|
"split_batch_2p5d",
|
|
"reduce_by_batch_2p5d",
|
|
"Linear2p5D",
|
|
"LayerNorm2p5D",
|
|
"Classifier2p5D",
|
|
"PatchEmbedding2p5D",
|
|
"Embedding2p5D",
|
|
"VocabParallelClassifier2p5D",
|
|
"VocabParallelEmbedding2p5D",
|
|
]
|