You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/shardformer/modeling
Wenxuan Tan 8fd25d6e09
[Feature] Split cross-entropy computation in SP (#5959)
3 months ago
..
chatglm2_6b
__init__.py
bert.py
blip2.py
bloom.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
chatglm2.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
command.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
deepseek.py
falcon.py
gpt2.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
gptj.py
jit.py
llama.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
mistral.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
mixtral.py
opt.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
qwen2.py [Feature] Split cross-entropy computation in SP (#5959) 3 months ago
sam.py
t5.py
vit.py
whisper.py