InternLM/internlm/core
ytxiong c219065348
feat(*): support sequence_parallel (#180)
* support sequence_parallel for no pipeline

* sequence_parallel does not support no-flash-attn

* support sequence parallel for pipeline

* add memory profiler

* Update 13B.py

* add memory profiler

* fix evaluation bug

* remove some unnecessary code

* remove some unnecessary code

* Update parallel_context.py

* modify the config

* remove memory profiler

* modify the config

* support selective dropout
2023-08-07 16:42:52 +08:00
..
communication refactor(scheduler): rewrite pipeline scheduler (#138) 2023-08-03 11:48:12 +08:00
context feat(*): support sequence_parallel (#180) 2023-08-07 16:42:52 +08:00
scheduler feat(*): support sequence_parallel (#180) 2023-08-07 16:42:52 +08:00
__init__.py initial commit 2023-07-06 12:55:23 +08:00
engine.py initial commit 2023-07-06 12:55:23 +08:00
gradient_handler.py initial commit 2023-07-06 12:55:23 +08:00
naive_amp.py refactor(scheduler): rewrite pipeline scheduler (#138) 2023-08-03 11:48:12 +08:00
trainer.py feat(core/scheduler): support pipeline parallel (#98) 2023-07-24 20:52:09 +08:00