Update README.md
Browse files
README.md
CHANGED
@@ -34,8 +34,9 @@ enhancing the fidelity of transcriptions in scenarios where punctuation is cruci
|
|
34 |
Finetuning dataset:** [MuST-C-en_ar](https://www.kaggle.com/datasets/sebaeymohamed/must-c-en-ar)
|
35 |
|
36 |
## Key Features:
|
37 |
-
Punctuation Sensitivity
|
38 |
accurate representation of the speaker's intended meaning.
|
|
|
39 |
|
40 |
|
41 |
## Usage
|
@@ -98,28 +99,28 @@ recordings from English TED Talks, which are automatically aligned at the senten
|
|
98 |
|Paramter|Value|
|
99 |
|-|-|
|
100 |
|per_device_train_batch_size|6|
|
101 |
-
|per_device_eval_batch_size|
|
102 |
-
|gradient_accumulation_steps|
|
103 |
|eval_accumulation_steps|16|
|
104 |
|dataloader_num_workers|2|
|
105 |
-
|learning_rate|
|
106 |
|adafactor|True|
|
107 |
-
|weight_decay|0.
|
108 |
-
|max_grad_norm|0.
|
109 |
-
|num_train_epochs|
|
110 |
-
|
|
111 |
|lr_scheduler_type|constant_with_warmup|
|
112 |
|fp16|True|
|
113 |
|gradient_checkpointing|True|
|
114 |
|sortish_sampler|True|
|
115 |
|
116 |
##### Results
|
117 |
-
**Train loss:** 0.
|
118 |
|Split|Word Error Rate (%)|
|
119 |
|-|-|
|
120 |
-
|dev|
|
121 |
-
|tst-HE|
|
122 |
-
|tst-COMMON|43.
|
123 |
|
124 |
|
125 |
## Citation
|
|
|
34 |
Finetuning dataset:** [MuST-C-en_ar](https://www.kaggle.com/datasets/sebaeymohamed/must-c-en-ar)
|
35 |
|
36 |
## Key Features:
|
37 |
+
**Punctuation Sensitivity:** The model is specifically engineered to be highly sensitive to punctuation nuances in spoken English, ensuring
|
38 |
accurate representation of the speaker's intended meaning.
|
39 |
+
**New Vocabulary:** Change vocabulary to be on Piece-level rather than character-level with vocabulary size 500 piece.
|
40 |
|
41 |
|
42 |
## Usage
|
|
|
99 |
|Paramter|Value|
|
100 |
|-|-|
|
101 |
|per_device_train_batch_size|6|
|
102 |
+
|per_device_eval_batch_size|16|
|
103 |
+
|gradient_accumulation_steps|12|
|
104 |
|eval_accumulation_steps|16|
|
105 |
|dataloader_num_workers|2|
|
106 |
+
|learning_rate|5e-5|
|
107 |
|adafactor|True|
|
108 |
+
|weight_decay|0.08989525|
|
109 |
+
|max_grad_norm|0.58585|
|
110 |
+
|num_train_epochs|5|
|
111 |
+
|warmup_ratio|0.7|
|
112 |
|lr_scheduler_type|constant_with_warmup|
|
113 |
|fp16|True|
|
114 |
|gradient_checkpointing|True|
|
115 |
|sortish_sampler|True|
|
116 |
|
117 |
##### Results
|
118 |
+
**Train loss:** 0.8925
|
119 |
|Split|Word Error Rate (%)|
|
120 |
|-|-|
|
121 |
+
|dev|44.8|
|
122 |
+
|tst-HE|39.1|
|
123 |
+
|tst-COMMON|43.2|
|
124 |
|
125 |
|
126 |
## Citation
|