You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/legacy/nn/layer/parallel_1d/__init__.py

26 lines
459 B

from .layers import (
Classifier1D,
Dropout1D,
Embedding1D,
LayerNorm1D,
Linear1D,
Linear1D_Col,
Linear1D_Row,
PatchEmbedding1D,
VocabParallelClassifier1D,
VocabParallelEmbedding1D,
)
__all__ = [
"Linear1D",
"Linear1D_Col",
"Linear1D_Row",
"Embedding1D",
"Dropout1D",
"Classifier1D",
"VocabParallelClassifier1D",
"VocabParallelEmbedding1D",
"LayerNorm1D",
"PatchEmbedding1D",
]