ColossalAI/colossalai/shardformer/layer
Edenzzzz 7ee569b05f
[hotfix] Fixed fused layernorm bug without apex (#5609)
* fixed fused layernorm bug without apex

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* same for flash attn

* remove flash attn check

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-04-24 23:04:06 +08:00
..
__init__.py [shardformer] refactor embedding resize (#5603) 2024-04-18 16:10:18 +08:00
_operation.py [shardformer] Sequence Parallelism Optimization (#5533) 2024-04-03 17:15:47 +08:00
attn.py [shardformer] update colo attention to support custom mask (#5510) 2024-03-27 11:19:32 +08:00
dropout.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
embedding.py [shardformer] refactor embedding resize (#5603) 2024-04-18 16:10:18 +08:00
linear.py [shardformer] refactor embedding resize (#5603) 2024-04-18 16:10:18 +08:00
loss.py [shardformer] refactor embedding resize (#5603) 2024-04-18 16:10:18 +08:00
normalization.py [hotfix] Fixed fused layernorm bug without apex (#5609) 2024-04-24 23:04:06 +08:00
parallel_module.py [shardformer] refactor embedding resize (#5603) 2024-04-18 16:10:18 +08:00
qkv_fused_linear.py [shardformer] Sequence Parallelism Optimization (#5533) 2024-04-03 17:15:47 +08:00
utils.py [shardformer] Sequence Parallelism Optimization (#5533) 2024-04-03 17:15:47 +08:00