ColossalAI/colossalai/shardformer/policies
Kun Lin 8af29ee47a [shardformer] support vision transformer (#4096)
* first v of vit shardformer

* keep vit

* update

* vit shard add vitattention vitlayer

* update num head shard para

* finish test for vit

* add new_model_class & postprocess

* add vit readme

* delete old files & fix the conflict

* fix sth
2023-07-04 16:05:01 +08:00
..
__init__.py [shardformer] init shardformer code structure (#3731) 2023-07-04 16:05:01 +08:00
autopolicy.py [shardformer] shardformer support opt models (#4091) 2023-07-04 16:05:01 +08:00
basepolicy.py [shardformer] refactored the shardformer layer structure (#4053) 2023-07-04 16:05:01 +08:00
bert.py [shardformer] support vision transformer (#4096) 2023-07-04 16:05:01 +08:00
gpt2.py [shardformer] add linearconv1d test (#4067) 2023-07-04 16:05:01 +08:00
llama.py [shardformer] adapted T5 and LLaMa test to use kit (#4049) 2023-07-04 16:05:01 +08:00
opt.py [shardformer] shardformer support opt models (#4091) 2023-07-04 16:05:01 +08:00
t5.py [shardformer] support vision transformer (#4096) 2023-07-04 16:05:01 +08:00
vit.py [shardformer] support vision transformer (#4096) 2023-07-04 16:05:01 +08:00