ColossalAI/extensions/inference
Steve Luo be396ad6cc
[Inference/Kernel] Add Paged Decoding kernel, sequence split within the same thread block (#5531)
* feat flash decoding for paged attention

* refactor flashdecodingattention

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-04-18 16:45:07 +08:00
..
__init__.py [Inference]Add CUDA KVCache Kernel (#5406) 2024-02-28 14:36:50 +08:00
inference_ops_cuda.py [Inference/Kernel] Add Paged Decoding kernel, sequence split within the same thread block (#5531) 2024-04-18 16:45:07 +08:00