ColossalAI/colossalai/inference/tensor_parallel
Bin Jia 1db6727678
[Pipeline inference] Combine kvcache with pipeline inference (#4938)
* merge kvcache with pipeline inference and refactor the code structure

* support ppsize > 2

* refactor pipeline code

* do pre-commit

* modify benchmark

* fix bench mark

* polish code

* add docstring and update readme

* refactor the code

* fix some logic bug of ppinfer

* polish readme

* fix typo

* skip infer test
2023-10-27 16:19:54 +08:00
..
modeling [Refactor] Integrated some lightllm kernels into token-attention (#4946) 2023-10-19 22:22:47 +08:00
policies [Refactor] Integrated some lightllm kernels into token-attention (#4946) 2023-10-19 22:22:47 +08:00
__init__.py [misc] update pre-commit and run all files (#4752) 2023-09-19 14:20:26 +08:00
batch_infer_state.py [Pipeline inference] Combine kvcache with pipeline inference (#4938) 2023-10-27 16:19:54 +08:00
engine.py [inference] add reference and fix some bugs (#4937) 2023-10-20 13:39:34 +08:00
kvcache_manager.py [Refactor] Integrated some lightllm kernels into token-attention (#4946) 2023-10-19 22:22:47 +08:00