ColossalAI/colossalai/kernel/triton
Hongxin Liu 641b1ee71a
[devops] remove post commit ci (#5566)
* [devops] remove post commit ci

* [misc] run pre-commit on all files

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-04-08 15:09:40 +08:00
..
__init__.py [Inference] Dynamic Batching Inference, online and offline (#4953) 2023-10-30 10:52:19 +08:00
context_attention.py [devops] remove post commit ci (#5566) 2024-04-08 15:09:40 +08:00
copy_kv_cache_dest.py [Inference] Dynamic Batching Inference, online and offline (#4953) 2023-10-30 10:52:19 +08:00
custom_autotune.py add autotune (#4822) 2023-09-28 13:47:35 +08:00
flash_decoding.py [devops] remove post commit ci (#5566) 2024-04-08 15:09:40 +08:00
fused_layernorm.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
gptq_triton.py [inference] add reference and fix some bugs (#4937) 2023-10-20 13:39:34 +08:00
int8_rotary_embedding_kernel.py [inference] Add smmoothquant for llama (#4904) 2023-10-16 11:28:44 +08:00
llama_act_combine_kernel.py [devops] remove post commit ci (#5566) 2024-04-08 15:09:40 +08:00
qkv_matmul_kernel.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
self_attention_nofusion.py [Refactor] Integrated some lightllm kernels into token-attention (#4946) 2023-10-19 22:22:47 +08:00
smooth_attention.py [inference] add reference and fix some bugs (#4937) 2023-10-20 13:39:34 +08:00
softmax.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
token_attention_kernel.py [devops] remove post commit ci (#5566) 2024-04-08 15:09:40 +08:00