ColossalAI/colossalai/utils
Jie Zhu 73d36618a6
[profiler] add MemProfiler (#356)
* add memory trainer hook

* fix bug

* add memory trainer hook

* fix import bug

* fix import bug

* add trainer hook

* fix #370 git log bug

* modify `to_tensorboard` function to support better output

* remove useless output

* change the name of `MemProfiler`

* complete memory profiler

* replace error with warning

* finish trainer hook

* modify interface of MemProfiler

* modify `__init__.py` in profiler

* remove unnecessary pass statement

* add usage to doc string

* add usage to trainer hook

* new location to store temp data file
2022-03-29 12:48:34 +08:00
..
data_sampler fixed utils docstring and add example to readme (#200) 2022-02-03 11:37:17 +08:00
gradient_accumulation Fixed docstring in colossalai (#171) 2022-01-21 10:44:30 +08:00
memory_tracer [profiler] add MemProfiler (#356) 2022-03-29 12:48:34 +08:00
memory_utils [zero] get memory usage of sharded optim v2. (#542) 2022-03-29 09:08:18 +08:00
multi_tensor_apply Fixed docstring in colossalai (#171) 2022-01-21 10:44:30 +08:00
profiler [profiler] add MemProfiler (#356) 2022-03-29 12:48:34 +08:00
tensor_detector Added tensor detector (#393) 2022-03-14 18:01:46 +08:00
__init__.py [memory] add model data tensor moving api (#503) 2022-03-24 14:29:41 +08:00
activation_checkpoint.py Added activation offload (#331) 2022-03-11 15:50:28 +08:00
checkpointing.py fixed mkdir conflict and align yapf config with flake (#220) 2022-02-15 11:31:13 +08:00
common.py [MOE] remove old MoE legacy (#493) 2022-03-22 17:37:16 +08:00
cuda.py Fixed docstring in colossalai (#171) 2022-01-21 10:44:30 +08:00
moe.py [polish] polish singleton and global context (#500) 2022-03-23 18:03:39 +08:00
timer.py [profiler] primary memory tracer 2022-03-11 15:50:28 +08:00