Xuanlei Zhao
|
dd2c28a323
|
[npu] use extension for op builder (#5172)
* update extension
* update cpu adam
* update is
* add doc for cpu adam
* update kernel
* update commit
* update flash
* update memory efficient
* update flash attn
* update flash attention loader
* update api
* fix
* update doc
* update example time limit
* reverse change
* fix doc
* remove useless kernel
* fix
* not use warning
* update
* update
|
2024-01-08 11:39:16 +08:00 |
Hongxin Liu
|
079bf3cb26
|
[misc] update pre-commit and run all files (#4752)
* [misc] update pre-commit
* [misc] run pre-commit
* [misc] remove useless configuration files
* [misc] ignore cuda for clang-format
|
2023-09-19 14:20:26 +08:00 |
flybird1111
|
7a3dfd0c64
|
[shardformer] update shardformer to use flash attention 2 (#4392)
* cherry-pick flash attention 2
cherry-pick flash attention 2
* [shardformer] update shardformer to use flash attention 2
[shardformer] update shardformer to use flash attention 2, fix
[shardformer] update shardformer to use flash attention 2, fix
[shardformer] update shardformer to use flash attention 2, fix
|
2023-08-15 23:25:14 +08:00 |
flybird1111
|
25c57b9fb4
|
[fix] coloattention support flash attention 2 (#4347)
Improved ColoAttention interface to support flash attention 2. Solved #4322
|
2023-08-04 13:46:22 +08:00 |
Frank Lee
|
dd14783f75
|
[kernel] fixed repeated loading of kernels (#2549)
* [kernel] fixed repeated loading of kernels
* polish code
* polish code
|
2023-02-03 09:47:13 +08:00 |
zbian
|
6877121377
|
updated flash attention api
|
2022-11-15 15:25:39 +08:00 |
ver217
|
f68eddfb3d
|
refactor kernel (#142)
|
2022-01-13 16:47:17 +08:00 |
shenggan
|
5c3843dc98
|
add colossalai kernel module (#55)
|
2021-12-21 12:19:52 +08:00 |