refactor toward model size
Browse files- run.sh → run-1b.sh +12 -10
- run-300M.sh +38 -0
run.sh → run-1b.sh
RENAMED
@@ -1,25 +1,26 @@
|
|
1 |
#!/bin/sh
|
2 |
|
3 |
-
export
|
|
|
4 |
|
5 |
python src/run_speech_recognition_ctc_bnb.py \
|
6 |
--dataset_name="mozilla-foundation/common_voice_7_0" \
|
7 |
--model_name_or_path="facebook/wav2vec2-xls-r-1b" \
|
8 |
--dataset_config_name="et" \
|
9 |
-
--output_dir="./" \
|
10 |
--overwrite_output_dir \
|
11 |
--num_train_epochs=100 \
|
12 |
-
--per_device_train_batch_size=
|
13 |
-
--per_device_eval_batch_size=
|
14 |
--gradient_accumulation_steps=2 \
|
15 |
--learning_rate=3e-4 \
|
16 |
-
--save_total_limit=
|
17 |
--warmup_steps=500 \
|
18 |
--evaluation_strategy=steps \
|
19 |
--text_column_name=sentence \
|
20 |
--length_column_name=input_length \
|
21 |
-
--save_steps=
|
22 |
-
--eval_steps=
|
23 |
--logging_steps=100 \
|
24 |
--layerdrop=0.0 \
|
25 |
--freeze_feature_encoder \
|
@@ -33,6 +34,7 @@ python src/run_speech_recognition_ctc_bnb.py \
|
|
33 |
--mask_time_length=10 \
|
34 |
--report_to=wandb \
|
35 |
--run_name="cosine+drop_proj+low_specaugment-1b" \
|
36 |
-
--do_train --do_eval
|
37 |
-
|
38 |
-
|
|
|
|
1 |
#!/bin/sh
|
2 |
|
3 |
+
export WANDB_PROJECT="xls-r-estonian"
|
4 |
+
export CUDA_VISIBLE_DEVICES=1
|
5 |
|
6 |
python src/run_speech_recognition_ctc_bnb.py \
|
7 |
--dataset_name="mozilla-foundation/common_voice_7_0" \
|
8 |
--model_name_or_path="facebook/wav2vec2-xls-r-1b" \
|
9 |
--dataset_config_name="et" \
|
10 |
+
--output_dir="./1B" \
|
11 |
--overwrite_output_dir \
|
12 |
--num_train_epochs=100 \
|
13 |
+
--per_device_train_batch_size=48 \
|
14 |
+
--per_device_eval_batch_size=48 \
|
15 |
--gradient_accumulation_steps=2 \
|
16 |
--learning_rate=3e-4 \
|
17 |
+
--save_total_limit=1 \
|
18 |
--warmup_steps=500 \
|
19 |
--evaluation_strategy=steps \
|
20 |
--text_column_name=sentence \
|
21 |
--length_column_name=input_length \
|
22 |
+
--save_steps=1000 \
|
23 |
+
--eval_steps=250 \
|
24 |
--logging_steps=100 \
|
25 |
--layerdrop=0.0 \
|
26 |
--freeze_feature_encoder \
|
|
|
34 |
--mask_time_length=10 \
|
35 |
--report_to=wandb \
|
36 |
--run_name="cosine+drop_proj+low_specaugment-1b" \
|
37 |
+
--do_train --do_eval
|
38 |
+
|
39 |
+
#--use_auth_token
|
40 |
+
#--push_to_hub
|
run-300M.sh
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#!/bin/sh
|
2 |
+
|
3 |
+
export WANDB_PROJECT="xls-r-estonian"
|
4 |
+
export CUDA_VISIBLE_DEVICES=2
|
5 |
+
|
6 |
+
python src/run_speech_recognition_ctc_bnb.py \
|
7 |
+
--dataset_name="mozilla-foundation/common_voice_7_0" \
|
8 |
+
--model_name_or_path="facebook/wav2vec2-xls-r-300m" \
|
9 |
+
--dataset_config_name="et" \
|
10 |
+
--output_dir="./" \
|
11 |
+
--overwrite_output_dir \
|
12 |
+
--num_train_epochs=100 \
|
13 |
+
--per_device_train_batch_size=80 \
|
14 |
+
--per_device_eval_batch_size=80 \
|
15 |
+
--gradient_accumulation_steps=2 \
|
16 |
+
--learning_rate=3e-4 \
|
17 |
+
--save_total_limit=1 \
|
18 |
+
--warmup_steps=500 \
|
19 |
+
--evaluation_strategy=steps \
|
20 |
+
--text_column_name=sentence \
|
21 |
+
--length_column_name=input_length \
|
22 |
+
--save_steps=1000 \
|
23 |
+
--eval_steps=250 \
|
24 |
+
--logging_steps=100 \
|
25 |
+
--layerdrop=0.0 \
|
26 |
+
--freeze_feature_encoder \
|
27 |
+
--feat_proj_dropout=0.1 \
|
28 |
+
--chars_to_ignore , ? . ! \- \; \: \" “ % ‘ ” � — ’ … – \
|
29 |
+
--gradient_checkpointing \
|
30 |
+
--lr_scheduler_type=cosine \
|
31 |
+
--fp16 \
|
32 |
+
--group_by_length \
|
33 |
+
--mask_time_prob=0.1 \
|
34 |
+
--mask_time_length=10 \
|
35 |
+
--report_to=wandb \
|
36 |
+
--run_name="cosine+drop_proj+low_specaugment-300M" \
|
37 |
+
--do_train --do_eval \
|
38 |
+
--use_auth_token --push_to_hub
|