12 Commits (ckpt)

Author SHA1 Message Date
flybird11111 6af6d6fc9f
[shardformer] support bias_gelu_jit_fused for models (#5647) 7 months ago
Wang Binluo 0d0a582033
[shardformer] update transformers (#5583) 7 months ago
Hongxin Liu 19e1a5cf16
[shardformer] update colo attention to support custom mask (#5510) 8 months ago
Frank Lee 7cfed5f076
[feat] refactored extension module (#5298) 10 months ago
Xuanlei Zhao dd2c28a323
[npu] use extension for op builder (#5172) 11 months ago
Hongxin Liu 1f5d2e8062
[hotfix] fix torch 2.0 compatibility (#4936) 1 year ago
Hongxin Liu 079bf3cb26
[misc] update pre-commit and run all files (#4752) 1 year ago
Baizhou Zhang 295b38fecf
[example] update vit example for hybrid parallel plugin (#4641) 1 year ago
Hongxin Liu 172f7fa3cf [misc] resolve code factor issues (#4433) 1 year ago
flybird1111 7a3dfd0c64 [shardformer] update shardformer to use flash attention 2 (#4392) 1 year ago
flybird1111 906426cb44 [Shardformer] Merge flash attention branch to pipeline branch (#4362) 1 year ago
FoolPlayer b3f5d7a3ba [shardformer] support pipeline base vit model (#4284) 1 year ago