ColossalAI/colossalai/shardformer/policies
Baizhou Zhang a14d352088 [pipeline] add pipeline forward for variants of gpt2 (#4238)
* add forward for GPTLMHeadModel

* add test for gpt_lm

* arranging get_held_layers method

* arrange forward replacement

* add forward for GPT2ForTokenClassification

* add forward for GPT2ForSequenceClassification

* fix test_shard_gpt2.py

* add GPT2DoubleHeadsmodel & fix bugs

* add id checking in get_shared_params
2023-08-15 23:25:14 +08:00
..
__init__.py [shardformer] init shardformer code structure (#3731) 2023-07-04 16:05:01 +08:00
auto_policy.py [pipeline] Llama pipeline (#4205) 2023-08-15 23:25:14 +08:00
base_policy.py [shardformer] fix base policy (#4229) 2023-08-15 23:25:14 +08:00
bert.py [pipeline] Llama pipeline (#4205) 2023-08-15 23:25:14 +08:00
bloom.py [pipeline] add bloom model pipeline (#4210) 2023-08-15 23:25:14 +08:00
gpt2.py [pipeline] add pipeline forward for variants of gpt2 (#4238) 2023-08-15 23:25:14 +08:00
llama.py [pipeline] Llama causal lm and llama for sequence classification pipeline (#4208) 2023-08-15 23:25:14 +08:00
opt.py [shardformer] support lazy init (#4202) 2023-08-15 23:25:14 +08:00
t5.py [shardformer] support lazy init (#4202) 2023-08-15 23:25:14 +08:00
vit.py [shardformer] rename policy file name 2023-08-15 23:25:14 +08:00