ColossalAI/colossalai/kernel/triton
Yuanheng Zhao e2c0e7f92a
[hotfix] Fix import error: colossal.kernel without triton installed (#4722)
* [hotfix] remove triton kernels from kernel init

* revise bloom/llama kernel imports for infer
2023-09-14 18:03:55 +08:00
..
__init__.py [hotfix] Fix import error: colossal.kernel without triton installed (#4722) 2023-09-14 18:03:55 +08:00
context_attention.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00
copy_kv_cache_dest.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00
fused_layernorm.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00
qkv_matmul_kernel.py [Kernels] added triton-implemented of self attention for colossal-ai (#4241) 2023-07-18 23:53:38 +08:00
rms_norm.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00
rotary_embedding_kernel.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00
self_attention_nofusion.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00
softmax.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00
token_attention_kernel.py [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) 2023-09-12 01:22:56 +08:00