InternLM/internlm/model
ytxiong 5ee651c2f1
feat(*): support not-flash-attn for pp and no-pp (#145)
* support not flash attention for no-pp

* support pipeline

* modify the config

* refactor the code

* refactor the code

* remove some unnecessary code
2023-07-28 16:13:04 +08:00
..
__init__.py feat(model/metrics.py): support calculating accuracy and perplexity m… (#91) 2023-07-26 16:22:10 +08:00
embedding.py [Dev] Pull Main (#139) 2023-07-27 10:20:21 +08:00
linear.py initial commit 2023-07-06 12:55:23 +08:00
loss.py initial commit 2023-07-06 12:55:23 +08:00
metrics.py feat(*): support not-flash-attn for pp and no-pp (#145) 2023-07-28 16:13:04 +08:00
modeling_internlm.py feat(*): support not-flash-attn for pp and no-pp (#145) 2023-07-28 16:13:04 +08:00
multi_head_attention.py feat(*): support not-flash-attn for pp and no-pp (#145) 2023-07-28 16:13:04 +08:00
utils.py initial commit 2023-07-06 12:55:23 +08:00