ColossalAI/colossalai/fx
Hongxin Liu b5f9e37c70
[legacy] clean up legacy code (#4743)
* [legacy] remove outdated codes of pipeline (#4692)

* [legacy] remove cli of benchmark and update optim (#4690)

* [legacy] remove cli of benchmark and update optim

* [doc] fix cli doc test

* [legacy] fix engine clip grad norm

* [legacy] remove outdated colo tensor (#4694)

* [legacy] remove outdated colo tensor

* [test] fix test import

* [legacy] move outdated zero to legacy (#4696)

* [legacy] clean up utils (#4700)

* [legacy] clean up utils

* [example] update examples

* [legacy] clean up amp

* [legacy] fix amp module

* [legacy] clean up gpc (#4742)

* [legacy] clean up context

* [legacy] clean core, constants and global vars

* [legacy] refactor initialize

* [example] fix examples ci

* [example] fix examples ci

* [legacy] fix tests

* [example] fix gpt example

* [example] fix examples ci

* [devops] fix ci installation

* [example] fix examples ci
2023-09-18 16:31:06 +08:00
..
codegen fix typo colossalai/auto_parallel autochunk fx/passes etc. (#3808) 2023-05-24 09:01:50 +08:00
passes [legacy] clean up legacy code (#4743) 2023-09-18 16:31:06 +08:00
profiler [hotfix] meta_tensor_compatibility_with_torch2 2023-03-30 13:43:01 +08:00
tracer [nfc] fix typo colossalai/cli fx kernel (#3847) 2023-06-02 15:02:45 +08:00
__init__.py [fx] metainfo_trace as an API. (#1873) 2022-11-10 20:58:37 +08:00
_compatibility.py [hotfix] meta_tensor_compatibility_with_torch2 2023-03-30 13:43:01 +08:00
_meta_regist_12.py [autoparallel] adapt autoparallel with new analyzer (#3261) 2023-03-30 17:47:24 +08:00
_meta_regist_13.py [fx] meta registration compatibility (#3253) 2023-03-27 15:22:17 +08:00
graph_module.py [fx] allow control of ckpt_codegen init (#2498) 2023-01-18 17:02:46 +08:00
proxy.py [NFC] policy colossalai/fx/proxy.py code style (#3269) 2023-03-29 15:22:21 +08:00