mirror of https://github.com/InternLM/InternLM
Update daily_tests.yaml
parent
202270a5e5
commit
89cabd477c
|
@ -36,12 +36,17 @@ jobs:
|
||||||
conda create -n internlm-model-latest --clone ${CONDA_BASE_ENV}
|
conda create -n internlm-model-latest --clone ${CONDA_BASE_ENV}
|
||||||
source activate internlm-model-latest
|
source activate internlm-model-latest
|
||||||
pip install transformers
|
pip install transformers
|
||||||
|
- name: install torch
|
||||||
|
run: |
|
||||||
|
source activate internlm-model-latest
|
||||||
|
pip install /mnt/petrelfs/qa-caif-cicd/resource/flash_attn-2.5.8+cu118torch2.2cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
||||||
|
pip install torch==2.2.2 torchvision==0.17.2 --index-url https://download.pytorch.org/whl/cu118
|
||||||
|
pip install sentencepiece auto-gptq==0.6.0 beautifulsoup4 decord lxml lmdeploy[all]
|
||||||
|
python3 -m pip list
|
||||||
|
conda deactivate
|
||||||
- name: run_test
|
- name: run_test
|
||||||
run: |
|
run: |
|
||||||
source activate internlm-model-latest
|
source activate internlm-model-latest
|
||||||
pip install torch==2.2.2 torchvision==0.17.2 --index-url https://download.pytorch.org/whl/cu118
|
|
||||||
pip install /mnt/petrelfs/qa-caif-cicd/resource/flash_attn-2.5.8+cu118torch2.2cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
|
||||||
pip install sentencepiece auto-gptq==0.6.0 beautifulsoup4 decord lxml
|
|
||||||
srun -p ${SLURM_PARTITION} --kill-on-bad-exit=1 --job-name=${GITHUB_RUN_ID}-${GITHUB_JOB} --gpus-per-task=2 pytest -s -v --color=yes ./tests/test_hf_model.py
|
srun -p ${SLURM_PARTITION} --kill-on-bad-exit=1 --job-name=${GITHUB_RUN_ID}-${GITHUB_JOB} --gpus-per-task=2 pytest -s -v --color=yes ./tests/test_hf_model.py
|
||||||
conda deactivate
|
conda deactivate
|
||||||
- name: remove_env
|
- name: remove_env
|
||||||
|
|
Loading…
Reference in New Issue