forked from mgrankin/ru_transformers
-
Notifications
You must be signed in to change notification settings - Fork 0
/
fit.sh
22 lines (22 loc) · 735 Bytes
/
fit.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
python tpu_lm_finetuning.py \
--seed=$RANDOM \
--output_dir=$OUTPUT \
--model_type=gpt2 \
--model_name_or_path=$MODEL_SIZE \
--do_train \
--train_data_file=$TRAIN_FILE \
--reload_data_file 1 \
--per_gpu_train_batch_size $BS \
--save_steps=10000 \
--logging_steps=100 \
--warmup_samples ${WARMUP:-128000} \
--learning_rate $LR \
--overwrite_output_dir \
--tokenizer_class YTEncoder \
--tokenizer_name bpe/yt.model \
--evaluate_during_training \
--eval_data_file=${VALID:-./data/classic/valid} \
--per_gpu_eval_batch_size $BS \
--save_total_limit 30 \
--num_train_epochs $NUM_EPOCH \
--unfreeze_level $UNFREEZE