ColossalAI/colossalai/kernel/npu
Xuanlei Zhao d6df19bae7
[npu] support triangle attention for llama (#5130)
* update fused attn

* update spda

* tri attn

* update triangle

* import

* fix

* fix
2023-11-30 14:21:30 +08:00
..
mha [npu] support triangle attention for llama (#5130) 2023-11-30 14:21:30 +08:00
__init__.py [npu] support triangle attention for llama (#5130) 2023-11-30 14:21:30 +08:00