Making large AI models cheaper, faster and more accessible
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
Wang Binluo 0d0a582033
[shardformer] update transformers (#5583)
7 months ago
..
__init__.py
_utils.py [shardformer] refactor embedding resize (#5603) 7 months ago
test_shard_bert.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago
test_shard_blip2.py [shardformer] update colo attention to support custom mask (#5510) 8 months ago
test_shard_bloom.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago
test_shard_chatglm2.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago
test_shard_falcon.py [devops] fix compatibility (#5444) 8 months ago
test_shard_gpt2.py [shardformer] Sequence Parallelism Optimization (#5533) 8 months ago
test_shard_gptj.py [shardformer] update colo attention to support custom mask (#5510) 8 months ago
test_shard_llama.py [shardformer] update transformers (#5583) 7 months ago
test_shard_mistral.py [shardformer] update transformers (#5583) 7 months ago
test_shard_opt.py [shardformer] update colo attention to support custom mask (#5510) 8 months ago
test_shard_sam.py [misc] update pre-commit and run all files (#4752) 1 year ago
test_shard_t5.py [shardformer] refactor embedding resize (#5603) 7 months ago
test_shard_vit.py [hotfix] fix torch 2.0 compatibility (#4936) 1 year ago
test_shard_whisper.py [shardformer] update transformers (#5583) 7 months ago