#!/bin/bash #argument1: model_path # launch server model_path=${1:-"lmsys/vicuna-7b-v1.3"} chat_template="{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}" echo "Model Path: $model_path" echo "Chat Tempelate" "${chat_template}" echo "Starting server..." python -m colossalai.inference.server.api_server --model $model_path --chat-template "${chat_template}" & SERVER_PID=$! # waiting time sleep 60 # Run Locust echo "Starting Locust..." echo "The test will automatically begin, you can turn to http://0.0.0.0:8089 for more information." echo "Test completion api first" locust -f locustfile.py -t 300 --tags online-generation --host http://127.0.0.1:8000 --autostart --users 300 --stop-timeout 10 echo "Test chat api" locust -f locustfile.py -t 300 --tags online-chat --host http://127.0.0.1:8000 --autostart --users 300 --stop-timeout 10 # kill Server echo "Stopping server..." kill $SERVER_PID echo "Test and server shutdown completely"