ColossalAI/colossalai/shardformer/layer
Frank Lee 58df720570 [shardformer] adapted T5 and LLaMa test to use kit (#4049)
* [shardformer] adapted T5 and LLaMa test to use kit

* polish code
2023-07-04 16:05:01 +08:00
..
__init__.py [shardformer] add gpt2 test and layer class refactor (#4041) 2023-07-04 16:05:01 +08:00
_operation.py [shardformer] integrated linear 1D with dtensor (#3996) 2023-07-04 16:05:01 +08:00
dist_crossentropy.py [shardformer] refactored embedding and dropout to parallel module (#4013) 2023-07-04 16:05:01 +08:00
dropout.py [shardformer] add gpt2 test and layer class refactor (#4041) 2023-07-04 16:05:01 +08:00
embedding1d.py [shardformer] adapted T5 and LLaMa test to use kit (#4049) 2023-07-04 16:05:01 +08:00
layernorm1d.py [shardformer] add gpt2 test and layer class refactor (#4041) 2023-07-04 16:05:01 +08:00
linear1d.py [shardformer] add gpt2 test and layer class refactor (#4041) 2023-07-04 16:05:01 +08:00
linearconv1d.py [shardformer] add gpt2 test and layer class refactor (#4041) 2023-07-04 16:05:01 +08:00
parallelmodule.py [shardformer] add gpt2 test and layer class refactor (#4041) 2023-07-04 16:05:01 +08:00
utils.py [shardformer] integrated linear 1D with dtensor (#3996) 2023-07-04 16:05:01 +08:00
vocabparallelembedding1d.py [shardformer] add gpt2 test and layer class refactor (#4041) 2023-07-04 16:05:01 +08:00