-
Notifications
You must be signed in to change notification settings - Fork 37
/
gpt3to4.sh
116 lines (104 loc) · 4.37 KB
/
gpt3to4.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
#!/usr/bin/env bash
# Copyright (c) Guangsheng Bao.
#
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
# setup the environment
echo `date`, Setup the environment ...
set -e # exit if error
# prepare folders
exp_path=exp_gpt3to4
data_path=$exp_path/data
res_path=$exp_path/results
mkdir -p $exp_path $data_path $res_path
datasets="xsum writing pubmed"
source_models="davinci gpt-3.5-turbo gpt-4"
# preparing dataset
openai_base="https://api.openai.com/v1"
openai_key="xxxxxxxx" # replace with your own key for generating your own test set
# We follow DetectGPT settings for generating text from GPT-3
M=davinci
for D in $datasets; do
echo `date`, Preparing dataset ${D} by sampling from openai/${M} ...
python scripts/data_builder.py --openai_model $M --openai_key $openai_key --openai_base $openai_base \
--dataset $D --n_samples 150 --do_top_p --top_p 0.9 --batch_size 1 \
--output_file $data_path/${D}_${M}
done
# We use a temperature of 0.8 for creativity writing
for M in gpt-3.5-turbo gpt-4; do
for D in $datasets; do
echo `date`, Preparing dataset ${D} by sampling from openai/${M} ...
python scripts/data_builder.py --openai_model $M --openai_key $openai_key --openai_base $openai_base \
--dataset $D --n_samples 150 --do_temperature --temperature 0.8 --batch_size 1 \
--output_file $data_path/${D}_${M}
done
done
# evaluate Fast-DetectGPT in the black-box setting
settings="gpt-j-6B:gpt2-xl gpt-j-6B:gpt-neo-2.7B gpt-j-6B:gpt-j-6B"
for M in $source_models; do
for D in $datasets; do
for S in $settings; do
IFS=':' read -r -a S <<< $S && M1=${S[0]} && M2=${S[1]}
echo `date`, Evaluating Fast-DetectGPT on ${D}_${M}.${M1}_${M2} ...
python scripts/fast_detect_gpt.py --reference_model_name $M1 --scoring_model_name $M2 --discrepancy_analytic \
--dataset $D --dataset_file $data_path/${D}_${M} --output_file $res_path/${D}_${M}.${M1}_${M2}
done
done
done
# evaluate supervised detectors
supervised_models="roberta-base-openai-detector roberta-large-openai-detector"
for M in $source_models; do
for D in $datasets; do
for SM in $supervised_models; do
echo `date`, Evaluating ${SM} on ${D}_${M} ...
python scripts/supervised.py --model_name $SM --dataset $D \
--dataset_file $data_path/${D}_${M} --output_file $res_path/${D}_${M}
done
done
done
# evaluate baselines
scoring_models="gpt-neo-2.7B"
for M in $source_models; do
for D in $datasets; do
for M2 in $scoring_models; do
echo `date`, Evaluating baseline methods on ${D}_${M}.${M2} ...
python scripts/baselines.py --scoring_model_name ${M2} --dataset $D \
--dataset_file $data_path/${D}_${M} --output_file $res_path/${D}_${M}.${M2}
done
done
done
# evaluate DNA-GPT
scoring_models="gpt-neo-2.7B"
for M in $source_models; do
for D in $datasets; do
for M2 in $scoring_models; do
echo `date`, Evaluating DNA-GPT on ${D}_${M}.${M2} ...
python scripts/dna_gpt.py --base_model_name ${M2} --dataset $D \
--dataset_file $data_path/${D}_${M} --output_file $res_path/${D}_${M}.${M2}
done
done
done
# evaluate DetectGPT and DetectLLM
scoring_models="gpt2-xl gpt-neo-2.7B gpt-j-6B"
for M in $source_models; do
for D in $datasets; do
M1=t5-11b # perturbation model
for M2 in $scoring_models; do
echo `date`, Evaluating DetectGPT on ${D}_${M}.${M1}_${M2} ...
python scripts/detect_gpt.py --mask_filling_model_name ${M1} --scoring_model_name ${M2} --n_perturbations 100 --dataset $D \
--dataset_file $data_path/${D}_${M} --output_file $res_path/${D}_${M}.${M1}_${M2}
# we leverage DetectGPT to generate the perturbations
echo `date`, Evaluating DetectLLM methods on ${D}_${M}.${M1}_${M2} ...
python scripts/detect_llm.py --scoring_model_name ${M2} --dataset $D \
--dataset_file $data_path/${D}_${M}.${M1}.perturbation_100 --output_file $res_path/${D}_${M}.${M1}_${M2}
done
done
done
# evaluate GPTZero
for M in $source_models; do
for D in $datasets; do
echo `date`, Evaluating GPTZero on ${D}_${M} ...
python scripts/gptzero.py --dataset $D \
--dataset_file $data_path/${D}_${M} --output_file $res_path/${D}_${M}
done
done