-
Notifications
You must be signed in to change notification settings - Fork 0
/
run_mlm.sh
43 lines (42 loc) · 1.4 KB
/
run_mlm.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
python scripts/run_mlm.py \
--model_name_or_path indobenchmark/indobert-base-p1 \
--tokenizer_name LazarusNLP/nusabert-base \
--max_seq_length 128 \
--per_device_train_batch_size 256 \
--per_device_eval_batch_size 256 \
--do_train --do_eval \
--max_steps 500000 \
--warmup_steps 24000 \
--learning_rate 3e-4 \
--weight_decay 0.01 \
--optim adamw_torch_fused \
--bf16 \
--preprocessing_num_workers 24 \
--dataloader_num_workers 24 \
--save_steps 10000 --save_total_limit 3 \
--output_dir outputs/nusabert-base \
--overwrite_output_dir \
--report_to tensorboard \
--push_to_hub --hub_private_repo \
--hub_model_id LazarusNLP/nusabert-base
python scripts/run_mlm.py \
--model_name_or_path indobenchmark/indobert-large-p1 \
--tokenizer_name LazarusNLP/NusaBERT-large \
--max_seq_length 128 \
--per_device_train_batch_size 256 \
--per_device_eval_batch_size 256 \
--do_train --do_eval \
--max_steps 500000 \
--warmup_steps 24000 \
--learning_rate 3e-5 \
--weight_decay 0.01 \
--optim adamw_torch_fused \
--bf16 \
--preprocessing_num_workers 24 \
--dataloader_num_workers 24 \
--save_steps 10000 --save_total_limit 3 \
--output_dir outputs/nusabert-large \
--overwrite_output_dir \
--report_to tensorboard \
--push_to_hub --hub_private_repo \
--hub_model_id LazarusNLP/NusaBERT-large