-
Notifications
You must be signed in to change notification settings - Fork 18
/
Copy pathllama-70b-urial.sh
32 lines (30 loc) · 1.04 KB
/
llama-70b-urial.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
version=$1
temp=${2:-0}
rp=${3:-1}
output_dir="result_dirs/mt-bench/urial_bench/"
mkdir -p $output_dir
gpu=0,1,2,3
n=4
CUDA_VISIBLE_DEVICES=$gpu python src/unified_infer.py \
--urial $version \
--download_dir /net/nfs/s2-research/llama2/ \
--model_name meta-llama/Llama-2-70b-hf \
--tensor_parallel_size $n \
--dtype bfloat16 \
--data_name mt-bench \
--mt_turn 1 \
--top_p 1 --temperature $temp --repetition_penalty $rp --batch_size 1 --max_tokens 2048 \
--filepath $output_dir/Llama-2-70b-hf.turn1.json \
--overwrite
CUDA_VISIBLE_DEVICES=$gpu python src/unified_infer.py \
--urial $version \
--download_dir /net/nfs/s2-research/llama2/ \
--model_name meta-llama/Llama-2-70b-hf \
--tensor_parallel_size $n \
--dtype bfloat16 \
--data_name mt-bench \
--mt_turn 2 \
--mt_turn1_result $output_dir/Llama-2-70b-hf.turn1.json \
--top_p 1 --temperature $temp --repetition_penalty $rp --batch_size 1 --max_tokens 2048 \
--filepath $output_dir/Llama-2-70b-hf.turn2.json \
--overwrite