Making large AI models cheaper, faster and more accessible
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
Wang Binluo dcd41d0973
Merge pull request #6071 from wangbluo/ring_attention
1 month ago
..
test_hybrid_parallel_grad_clip_norm [MoE/ZeRO] Moe refactor with zero refactor (#5821) 5 months ago
test_layer Merge pull request #6071 from wangbluo/ring_attention 1 month ago
test_model [moe] add parallel strategy for shared_expert && fix test for deepseek (#6063) 2 months ago
__init__.py [shardformer] adapted T5 and LLaMa test to use kit (#4049) 1 year ago
test_flash_attention.py [Feature] Zigzag Ring attention (#5905) 3 months ago
test_shard_utils.py [misc] update pre-commit and run all files (#4752) 1 year ago
test_with_torch_ddp.py [misc] refactor launch API and tensor constructor (#5666) 7 months ago