-
Notifications
You must be signed in to change notification settings - Fork 1
/
dist_run.sh
102 lines (88 loc) · 3.42 KB
/
dist_run.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
#!/bin/bash
if [ $# -lt 3 ]; then
echo "USAGE: $0 DATASET METHOD CONFIG ..."
exit
fi
WORK_DIR='work_dirs'
DATASET=$1
METHOD=$2 # random, coreset, mean_entropy, miaod etc
CONFIG=$3
BACKBONE='faster'
GPUS=${GPUS:-4}
IMGS_PER_GPU=${IMGS_PER_GPU:-4}
PORT=${PORT:-29504}
TRAIN_STEP=${TRAIN_STEP:-10}
QUERY_UNIT=${QUERY_UNIT:-'box'}
INIT_NUM=${INIT_NUM:-2000}
ADD_NUM=${ADD_NUM:-1000}
DELETE_MODEL=${DELETE_MODEL:-0}
seed=${SEED:-2022}
START_ITER=${START_ITER:-0}
SUFFIX=""
if [ ${QUERY_UNIT} == 'box' ]; then
SUFFIX="--count-box ${SUFFIX}"
fi
if [ ${METHOD} == 'miaod' ]; then
SUFFIX="--training-setting semi ${SUFFIX}"
fi
export PYTHONPATH="$(dirname $0)":$PYTHONPATH
branch_name=`git branch --show-current`
short_sha=`git rev-parse --short=11 HEAD`
unique_arch_list="learning_loss almdn miaod mcdropout"
# else random mean_entropy wbqbc boxcnt
epochs=""
if [[ ${@:4} == *"epochs"* ]]; then
tmpvar=${@:4}
tmpvar=${tmpvar#*epochs}
epochs="_e`echo ${tmpvar%--*} | sed 's/ //g'`"
fi
TEMPLATE=${DATASET}_${BACKBONE}_${QUERY_UNIT}_${INIT_NUM}_${ADD_NUM}${epochs}_template_${seed}_${branch_name}_${short_sha}
TIMESTAMP=${DATASET}_${BACKBONE}_${QUERY_UNIT}_${INIT_NUM}_${ADD_NUM}${epochs}_${METHOD}_$(date "+%Y%m%d%H%M%S")_${seed}_${branch_name}_${short_sha}
if [ -d "${WORK_DIR}/$TEMPLATE" ]; then
echo "Copying from $TEMPLATE to $TIMESTAMP ..."
cp -rdT ${WORK_DIR}/${TEMPLATE} ${WORK_DIR}/${TIMESTAMP}
[[ " $unique_arch_list " =~ " $METHOD " ]] && train_ini=true || train_ini=false
else
mkdir -p ${WORK_DIR}/${TIMESTAMP}
train_ini=true
fi
for ((i=${START_ITER};i<${TRAIN_STEP};i++))
do
workdir=${WORK_DIR}/${TIMESTAMP}/step_${i}
mkdir -p ${workdir}
if [ ${i} == 0 ]
then
if ${train_ini}
then
rm ${workdir}/*.pth
python -m torch.distributed.launch --nproc_per_node=$GPUS --master_port=$PORT \
$(dirname "$0")/tools/train.py $CONFIG --initial-ratio=${INIT_NUM} \
--cfg-options data.samples_per_gpu=${IMGS_PER_GPU} data.workers_per_gpu=${IMGS_PER_GPU} \
--work-dir=${workdir} --ratio=0 --seed=$seed \
--mining-method=${METHOD} ${SUFFIX} --launcher pytorch ${@:4}
fi
else
j=`expr ${i} - 1`
prev_workdir=${WORK_DIR}/${TIMESTAMP}/step_${j}
python -m torch.distributed.launch --nproc_per_node=$GPUS --master_port=$PORT \
$(dirname "$0")/tools/train.py $CONFIG \
--model-result=${prev_workdir}/results.bbox.json \
--cfg-options data.samples_per_gpu=${IMGS_PER_GPU} data.workers_per_gpu=${IMGS_PER_GPU} \
data.train.labeled.ann_file=${prev_workdir}/labeled.json \
data.train.unlabeled.ann_file=${prev_workdir}/unlabeled.json \
--work-dir=${workdir} --ratio=${ADD_NUM} --seed=$seed \
--mining-method=${METHOD} ${SUFFIX} --launcher pytorch ${@:4}
fi
# skip the last ratio(no data in unlabled annotation file)
if [ ${i} != `expr ${TRAIN_STEP} - 1` ]; then
ckpt="$(ls ${workdir}/*.pth | sort -V | tail -n1)"
python -m torch.distributed.launch --nproc_per_node=$GPUS --master_port=$PORT \
$(dirname "$0")/tools/test.py $CONFIG ${ckpt} \
--mining-method=${METHOD} --seed=$seed --work-dir=${workdir} \
--cfg-options data.test.ann_file=${workdir}/unlabeled.json \
--out ${workdir}/results.pkl --format-only --gpu-collect --launcher pytorch
fi
done
if [ ${DELETE_MODEL} == 1 ]; then
find ${WORK_DIR}/${TIMESTAMP} -name '*.pth' -type f -print -exec rm -rf {} \;
fi