ColossalAI/colossalai/inference/tensor_parallel/modeling
Jianghai ef4c14a5e2
[Inference] Fix bug in ChatGLM2 Tensor Parallelism (#5014)
* fix bug

* fix

* fix multiquery

* fix multiquery

---------

Co-authored-by: CjhHa1 <cjh18671720497outlook.com>
2023-11-07 15:01:50 +08:00
..
__init__.py [inference]fix import bug and delete down useless init (#4830) 2023-10-04 09:18:45 +08:00
_utils.py [inference] add llama2 support (#4898) 2023-10-13 13:09:23 +08:00
bloom.py [Kernels]Updated Triton kernels into 2.1.0 and adding flash-decoding for llama token attention (#4965) 2023-10-30 14:04:37 +08:00
chatglm2.py [Inference] Fix bug in ChatGLM2 Tensor Parallelism (#5014) 2023-11-07 15:01:50 +08:00
llama.py [Pipeline Inference] Merge pp with tp (#4993) 2023-11-01 12:46:21 +08:00