ColossalAI/tests/test_infer_ops
Cuiqing Li e937461312
[Bug FIx] import llama context ops fix (#4524)
* added _vllm_rms_norm

* change place

* added tests

* added tests

* modify

* adding kernels

* added tests:

* adding kernels

* modify

* added

* updating kernels

* adding tests

* added tests

* kernel change

* submit

* modify

* added

* edit comments

* change name

* change commnets and fix import

* add

* added

* fix

* add ops into init.py

* add
2023-08-28 13:41:57 +08:00
..
cuda [Kernels] add necessary kernels (llama & bloom) for attention forward and kv-cache manager (#4485) 2023-08-24 16:30:02 +08:00
triton [Bug FIx] import llama context ops fix (#4524) 2023-08-28 13:41:57 +08:00