ColossalAI/tests/test_infer_ops/triton
Cuiqing Li (李崔卿) 28052a71fb
[Kernels]Update triton kernels into 2.1.0 (#5046)
* update flash-context-attention

* adding kernels

* fix

* reset

* add build script

* add building process

* add llama2 exmaple

* add colossal-llama2 test

* clean

* fall back test setting

* fix test file

* clean

* clean

* clean

---------

Co-authored-by: cuiqing.li <lixx336@gmail.com>
2023-11-16 16:43:15 +08:00
..
kernel_utils.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
test_bloom_context_attention.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
test_copy_kv_dest.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
test_layernorm_triton.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
test_llama_act_combine.py [moe] merge moe into main (#4978) 2023-11-02 02:21:24 +00:00
test_llama_context_attention.py [Kernels]Update triton kernels into 2.1.0 (#5046) 2023-11-16 16:43:15 +08:00
test_self_attention_nonfusion.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
test_softmax.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
test_token_attn_fwd.py [Refactor] Integrated some lightllm kernels into token-attention (#4946) 2023-10-19 22:22:47 +08:00
test_token_softmax.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00