mirror of https://github.com/InternLM/InternLM
[CI]: Delete .github/workflows/demo_in_readme.yaml (#632)
parent
6e51133a4d
commit
519c7934c4
|
@ -1,128 +0,0 @@
|
||||||
name: demo-in-readme
|
|
||||||
on:
|
|
||||||
pull_request:
|
|
||||||
branches:
|
|
||||||
- "main"
|
|
||||||
- "develop"
|
|
||||||
paths-ignore:
|
|
||||||
- "docs/**"
|
|
||||||
- "**.md"
|
|
||||||
env:
|
|
||||||
WORKSPACE_PREFIX: $(echo $GITHUB_WORKSPACE |cut -d '/' -f 1-4)
|
|
||||||
SLURM_PARTITION: llm_s
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
check-requirements:
|
|
||||||
runs-on: [t_cluster]
|
|
||||||
steps:
|
|
||||||
- name: mask env
|
|
||||||
run: |
|
|
||||||
echo "::add-mask::${{env.WORKSPACE_PREFIX}}"
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
with:
|
|
||||||
fetch-depth: 2
|
|
||||||
- name: check-requirements
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
changed_files=$(git diff --name-only -r HEAD^1 HEAD)
|
|
||||||
echo $changed_files
|
|
||||||
if [[ $changed_files =~ "runtime.txt" ]]; then
|
|
||||||
pip install -r requirements/runtime.txt
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [[ $changed_files =~ "torch.txt" ]]; then
|
|
||||||
pip install -r requirements/torch.txt
|
|
||||||
fi
|
|
||||||
|
|
||||||
dataset-preparation:
|
|
||||||
if: ${{ always() }}
|
|
||||||
needs: check-requirements
|
|
||||||
runs-on: [t_cluster]
|
|
||||||
steps:
|
|
||||||
- name: mask env
|
|
||||||
run: |
|
|
||||||
echo "::add-mask::${{env.WORKSPACE_PREFIX}}"
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
|
|
||||||
- name: raw-chinese-data
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
sh ./ci_scripts/data/tokenizer_chinese.sh ${GITHUB_RUN_ID}-${GITHUB_JOB}
|
|
||||||
|
|
||||||
- name: alpaca-data
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
sh ./ci_scripts/data/tokenizer_alpaca.sh
|
|
||||||
|
|
||||||
train:
|
|
||||||
if: ${{ always() }}
|
|
||||||
needs: check-requirements
|
|
||||||
runs-on: [t_cluster]
|
|
||||||
timeout-minutes: 30
|
|
||||||
steps:
|
|
||||||
- name: mask env
|
|
||||||
run: |
|
|
||||||
echo "::add-mask::${{env.WORKSPACE_PREFIX}}"
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
|
|
||||||
- name: slurm-train
|
|
||||||
id: basic_train
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
sh ./ci_scripts/train/slurm_train.sh ${GITHUB_RUN_ID}-${GITHUB_JOB}
|
|
||||||
|
|
||||||
- name: load_preset_ckpt
|
|
||||||
if: ${{ failure() && steps.basic_train.conclusion == 'failure' }}
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
export PYTHONPATH=$PWD:$PYTHONPATH
|
|
||||||
sh ./ci_scripts/train/load_ckpt.sh 7B_load_preset_ckpt ${GITHUB_RUN_ID}-${GITHUB_JOB}
|
|
||||||
|
|
||||||
- name: load_new_ckpt
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
export PYTHONPATH=$PWD:$PYTHONPATH
|
|
||||||
sh ./ci_scripts/train/load_ckpt.sh 7B_load_new_ckpt ${GITHUB_RUN_ID}-${GITHUB_JOB}
|
|
||||||
rsync -av --remove-source-files $GITHUB_WORKSPACE/llm_ckpts ${{env.WORKSPACE_PREFIX}}/ci_clean_bak
|
|
||||||
|
|
||||||
- name: torchrun-train
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
sh ./ci_scripts/train/torchrun.sh ${GITHUB_RUN_ID}-${GITHUB_JOB}
|
|
||||||
rsync -av --remove-source-files $GITHUB_WORKSPACE/llm_ckpts ${{env.WORKSPACE_PREFIX}}/ci_clean_bak
|
|
||||||
|
|
||||||
convert-model-then-load:
|
|
||||||
if: ${{ always() }}
|
|
||||||
needs: check-requirements
|
|
||||||
runs-on: [t_cluster]
|
|
||||||
timeout-minutes: 15
|
|
||||||
steps:
|
|
||||||
- name: mask env
|
|
||||||
run: |
|
|
||||||
echo "::add-mask::${{env.WORKSPACE_PREFIX}}"
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
|
|
||||||
- name: convert-model-then-load
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
export PYTHONPATH=$PWD:$PYTHONPATH
|
|
||||||
sh ./ci_scripts/model/convert_to_hf.sh
|
|
||||||
cd ./hf_ckpt
|
|
||||||
srun -p ${SLURM_PARTITION} --quotatype=spot --job-name=${GITHUB_RUN_ID}-${GITHUB_JOB} --gpus-per-task=2 python ../ci_scripts/model/loaded_as_transformer.py
|
|
||||||
cd ..
|
|
||||||
rsync -av --remove-source-files $GITHUB_WORKSPACE/hf_ckpt ${{env.WORKSPACE_PREFIX}}/ci_clean_bak
|
|
||||||
load-chat-model-in-hf:
|
|
||||||
if: ${{ always() }}
|
|
||||||
needs: check-requirements
|
|
||||||
runs-on: [t_cluster]
|
|
||||||
timeout-minutes: 15
|
|
||||||
steps:
|
|
||||||
- name: mask env
|
|
||||||
run: |
|
|
||||||
echo "::add-mask::${{env.WORKSPACE_PREFIX}}"
|
|
||||||
- uses: actions/checkout@v3
|
|
||||||
|
|
||||||
- name: chat-model-in-hf
|
|
||||||
run: |
|
|
||||||
source activate internlm-env-test
|
|
||||||
srun -p ${SLURM_PARTITION} --quotatype=spot --job-name=${GITHUB_RUN_ID}-${GITHUB_JOB} --gpus-per-task=2 python ./ci_scripts/model/demo_load_7B_chat_model.py
|
|
Loading…
Reference in New Issue