You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/shardformer/modeling
flybird11111 c7d6975d29
[shardformer] fix GPT2DoubleHeadsModel (#4703)
1 year ago
..
chatglm2_6b [pipeline] add chatglm (#4363) 1 year ago
__init__.py [shardformer] added development protocol for standardization (#4149) 1 year ago
bert.py [shardformer] vit/llama/t5 ignore the sequence parallelism flag and some fix. (#4498) 1 year ago
blip2.py [shardformer] update shardformer to use flash attention 2 (#4392) 1 year ago
bloom.py [shardformer] bloom support sequence parallel (#4465) 1 year ago
chatglm2.py [shardformer] chatglm support sequence parallel (#4482) 1 year ago
gpt2.py [shardformer] fix GPT2DoubleHeadsModel (#4703) 1 year ago
jit.py [Shardformer] Merge flash attention branch to pipeline branch (#4362) 1 year ago
llama.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 1 year ago
opt.py [shardformer] update llama2/opt finetune example and fix llama2 policy (#4645) 1 year ago
sam.py [Shardformer] Merge flash attention branch to pipeline branch (#4362) 1 year ago
t5.py [misc] resolve code factor issues (#4433) 1 year ago
vit.py [example] update vit example for hybrid parallel plugin (#4641) 1 year ago
whisper.py [shardformer] Pipeline/whisper (#4456) 1 year ago