Update directory structure for NPU-related files
- Create a new `npu` directory under `assets` to store NPU-related images. - Move NPU-related `README` files to the `ecosystem` directory.pull/816/head
Before Width: | Height: | Size: 363 KiB After Width: | Height: | Size: 363 KiB |
Before Width: | Height: | Size: 41 KiB After Width: | Height: | Size: 41 KiB |
Before Width: | Height: | Size: 8.9 KiB After Width: | Height: | Size: 8.9 KiB |
Before Width: | Height: | Size: 212 KiB After Width: | Height: | Size: 212 KiB |
Before Width: | Height: | Size: 6.9 KiB After Width: | Height: | Size: 6.9 KiB |
Before Width: | Height: | Size: 315 KiB After Width: | Height: | Size: 315 KiB |
|
@ -14,8 +14,8 @@
|
|||
<div> </div>
|
||||
</div>
|
||||
|
||||
[](./LICENSE)
|
||||
[](https://github.com/internLM/OpenCompass/)
|
||||
[](../LICENSE)
|
||||
[](https://github.com/internLM/OpenCompass/)
|
||||
|
||||
<!-- [](https://internlm.readthedocs.io/zh_CN/latest/?badge=latest) -->
|
||||
|
||||
|
@ -28,8 +28,8 @@
|
|||
[🔗API](https://internlm.intern-ai.org.cn/api/document) |
|
||||
[🧩Modelers](https://modelers.cn/spaces/MindSpore-Lab/INTERNLM2-20B-PLAN)
|
||||
|
||||
[English](./README_npu.md) |
|
||||
[简体中文](./README_npu_zh-CN.md)
|
||||
[English](README_npu.md) |
|
||||
[简体中文](README_npu_zh-CN.md)
|
||||
|
||||
</div>
|
||||
|
||||
|
@ -140,7 +140,7 @@ NPROC_PER_NODE=8 xtuner train internlm3_8b_instruct_lora_oasst1_e10.py --deepspe
|
|||
The fine-tuning results are saved in the directory `./work_dirs/internlm3_8b_instruct_lora_oasst1_e10/iter_xxx.pth`.
|
||||
The comparison of loss between NPU and GPU is as follows:
|
||||
|
||||

|
||||

|
||||
|
||||
### Model Convert
|
||||
|
||||
|
@ -254,11 +254,11 @@ llamafactory-cli train examples/train_full/internlm3_8b_instruct_full_sft.yaml
|
|||
|
||||
The loss curve obtained after finetuning is as follows:
|
||||
|
||||

|
||||

|
||||
|
||||
The loss curve compared with GPU is as follows:
|
||||
|
||||

|
||||

|
||||
|
||||
## Transformers
|
||||
|
|
@ -28,8 +28,8 @@
|
|||
[🔗API](https://internlm.intern-ai.org.cn/api/document) |
|
||||
[🧩魔乐社区](https://modelers.cn/spaces/MindSpore-Lab/INTERNLM2-20B-PLAN)
|
||||
|
||||
[English](./README_npu.md) |
|
||||
[简体中文](./README_npu_zh-CN.md)
|
||||
[English](README_npu.md) |
|
||||
[简体中文](README_npu_zh-CN.md)
|
||||
|
||||
</div>
|
||||
|
||||
|
@ -139,7 +139,7 @@ NPROC_PER_NODE=8 xtuner train internlm3_8b_instruct_lora_oasst1_e10.py --deepspe
|
|||
|
||||
微调后结果保存在`./work_dirs/internlm3_8b_instruct_lora_oasst1_e10/iter_xxx.pth`,NPU与GPU的loss对比如下:
|
||||
|
||||

|
||||

|
||||
|
||||
### 模型转换
|
||||
|
||||
|
@ -250,11 +250,11 @@ llamafactory-cli train examples/train_full/internlm3_8b_instruct_full_sft.yaml
|
|||
|
||||
微调后得到的loss曲线如下:
|
||||
|
||||

|
||||

|
||||
|
||||
与GPU对比的loss曲线如下:
|
||||
|
||||

|
||||

|
||||
|
||||
## Transformers
|
||||
|