ColossalAI/colossalai/inference/modeling
yuehuayingxueluo 3c91e3f176
[Inference]Adapt to baichuan2 13B (#5614)
* adapt to baichuan2 13B

* adapt to baichuan2 13B

* change BAICHUAN_MODEL_NAME_OR_PATH

* fix test_decoding_attn.py

* Modifications based on review comments.

* change BAICHUAN_MODEL_NAME_OR_PATH

* mv attn mask processes to test flash decoding

* mv get_alibi_slopes baichuan modeling

* fix bugs in test_baichuan.py
2024-04-25 23:11:30 +08:00
..
layers [Inference]Fused the gate and up proj in mlp,and optimized the autograd process. (#5365) 2024-02-06 19:38:25 +08:00
models [Inference]Adapt to baichuan2 13B (#5614) 2024-04-25 23:11:30 +08:00
policy [Inference]Adapt to baichuan2 13B (#5614) 2024-04-25 23:11:30 +08:00
__init__.py [doc] updated inference readme (#5343) 2024-02-02 14:31:10 +08:00