You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/tests/test_shardformer/test_model
littsk 1a3315e336
[hotfix] Add layer norm gradients all-reduce for sequence parallel (#4926)
1 year ago
..
__init__.py [shardformer] adapted T5 and LLaMa test to use kit (#4049) 1 year ago
_utils.py [hotfix] fix torch 2.0 compatibility (#4936) 1 year ago
test_shard_bert.py [hotfix] Add layer norm gradients all-reduce for sequence parallel (#4926) 1 year ago
test_shard_blip2.py [misc] update pre-commit and run all files (#4752) 1 year ago
test_shard_bloom.py [hotfix] Add layer norm gradients all-reduce for sequence parallel (#4926) 1 year ago
test_shard_chatglm2.py [hotfix] Add layer norm gradients all-reduce for sequence parallel (#4926) 1 year ago
test_shard_gpt2.py [hotfix] Add layer norm gradients all-reduce for sequence parallel (#4926) 1 year ago
test_shard_llama.py [misc] update pre-commit and run all files (#4752) 1 year ago
test_shard_opt.py [misc] update pre-commit and run all files (#4752) 1 year ago
test_shard_sam.py [misc] update pre-commit and run all files (#4752) 1 year ago
test_shard_t5.py [misc] update pre-commit and run all files (#4752) 1 year ago
test_shard_vit.py [hotfix] fix torch 2.0 compatibility (#4936) 1 year ago
test_shard_whisper.py [misc] update pre-commit and run all files (#4752) 1 year ago