-
Notifications
You must be signed in to change notification settings - Fork 12
Expand file tree
/
Copy pathrun.sh
More file actions
executable file
·30 lines (24 loc) · 978 Bytes
/
run.sh
File metadata and controls
executable file
·30 lines (24 loc) · 978 Bytes
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
#!/bin/bash
set -e # Exit immediately on error.
if [ $# -lt 1 ] || [ $# -gt 2 ]; then
echo "Usage: $0 <model_location> <agent>"
echo "Example: $0 Qwen/Qwen2.5-7B-Instruct"
echo "Note: agent parameter is accepted for consistency but not used by this benchmark"
exit 1
fi
MODEL_NAME="$1"
NEW_MODEL_NAME="${MODEL_NAME//\//_}"
# Note: set it to "openai" if you are using your own model server (vllm)
# Otherwise, set it to "azure" if you are using azure gpt endpoint
# Run self-serving model
# export OPENAI_API_TYPE="openai"
# export OPENAI_BASE_URL="http://localhost:2327/v1"
# export OPENAI_API_KEY="EMPTY"
source .venv/bin/activate
echo "==> Start to run ExampleBench"
# Note that if you benchmark has multiple tasks, you need to add --task <task>
# in your code to enable task selection.
python src/main.py \
--model_name "${MODEL_NAME}"
# --save_path "./outputs/examplebench__${NEW_MODEL_NAME}__$(date +"%Y-%m-%d_%H-%M-%S")" \
deactivate