Hongxin Liu
|
19e1a5cf16
|
[shardformer] update colo attention to support custom mask (#5510)
* [feature] refactor colo attention (#5462)
* [extension] update api
* [feature] add colo attention
* [feature] update sdpa
* [feature] update npu attention
* [feature] update flash-attn
* [test] add flash attn test
* [test] update flash attn test
* [shardformer] update modeling to fit colo attention (#5465)
* [misc] refactor folder structure
* [shardformer] update llama flash-attn
* [shardformer] fix llama policy
* [devops] update tensornvme install
* [test] update llama test
* [shardformer] update colo attn kernel dispatch
* [shardformer] update blip2
* [shardformer] update chatglm
* [shardformer] update gpt2
* [shardformer] update gptj
* [shardformer] update opt
* [shardformer] update vit
* [shardformer] update colo attention mask prep
* [shardformer] update whisper
* [test] fix shardformer tests (#5514)
* [test] fix shardformer tests
* [test] fix shardformer tests
|
2024-03-27 11:19:32 +08:00 |
Hongxin Liu
|
070df689e6
|
[devops] fix extention building (#5427)
|
2024-03-05 15:35:54 +08:00 |
Hongxin Liu
|
ffffc32dc7
|
[checkpointio] fix gemini and hybrid parallel optim checkpoint (#5347)
* [checkpointio] fix hybrid parallel optim checkpoint
* [extension] fix cuda extension
* [checkpointio] fix gemini optimizer checkpoint
* polish code
|
2024-02-01 16:13:06 +08:00 |
Frank Lee
|
abd8e77ad8
|
[extension] fixed exception catch (#5342)
|
2024-01-31 18:09:49 +08:00 |
digger yu
|
6a3086a505
|
fix typo under extensions/ (#5330)
|
2024-01-30 09:55:16 +08:00 |
Frank Lee
|
febed23288
|
[doc] added docs for extensions (#5324)
* [doc] added docs for extensions
* polish
* polish
|
2024-01-29 17:39:23 +08:00 |
Frank Lee
|
7cfed5f076
|
[feat] refactored extension module (#5298)
* [feat] refactored extension module
* polish
* polish
* polish
* polish
* polish
* polish
* polish
* polish
* polish
* polish
|
2024-01-25 17:01:48 +08:00 |