GPUS=$1 | |
CONFIG=$2 | |
WORK_DIR=$3 | |
CHECKPOINT=$4 | |
PORT=${PORT:-29500} | |
python -m torch.distributed.launch \ | |
--nproc_per_node=${GPUS} \ | |
--master_port=${PORT} \ | |
./test.py ${CONFIG} \ | |
--work_dir ${WORK_DIR} \ | |
--checkpoint ${CHECKPOINT} \ | |
--launcher="pytorch" \ | |
${@:5} | |