seba3y commited on
Commit
54f0bb6
·
1 Parent(s): 40ea9c1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -12
README.md CHANGED
@@ -34,8 +34,9 @@ enhancing the fidelity of transcriptions in scenarios where punctuation is cruci
34
  Finetuning dataset:** [MuST-C-en_ar](https://www.kaggle.com/datasets/sebaeymohamed/must-c-en-ar)
35
 
36
  ## Key Features:
37
- Punctuation Sensitivity: The model is specifically engineered to be highly sensitive to punctuation nuances in spoken English, ensuring
38
  accurate representation of the speaker's intended meaning.
 
39
 
40
 
41
  ## Usage
@@ -98,28 +99,28 @@ recordings from English TED Talks, which are automatically aligned at the senten
98
  |Paramter|Value|
99
  |-|-|
100
  |per_device_train_batch_size|6|
101
- |per_device_eval_batch_size|10|
102
- |gradient_accumulation_steps|20|
103
  |eval_accumulation_steps|16|
104
  |dataloader_num_workers|2|
105
- |learning_rate|7e-5|
106
  |adafactor|True|
107
- |weight_decay|0.1|
108
- |max_grad_norm|0.9|
109
- |num_train_epochs|2.15|
110
- |warmup_steps|2000|
111
  |lr_scheduler_type|constant_with_warmup|
112
  |fp16|True|
113
  |gradient_checkpointing|True|
114
  |sortish_sampler|True|
115
 
116
  ##### Results
117
- **Train loss:** 0.4429
118
  |Split|Word Error Rate (%)|
119
  |-|-|
120
- |dev|51.6|
121
- |tst-HE|40.2|
122
- |tst-COMMON|43.01|
123
 
124
 
125
  ## Citation
 
34
  Finetuning dataset:** [MuST-C-en_ar](https://www.kaggle.com/datasets/sebaeymohamed/must-c-en-ar)
35
 
36
  ## Key Features:
37
+ **Punctuation Sensitivity:** The model is specifically engineered to be highly sensitive to punctuation nuances in spoken English, ensuring
38
  accurate representation of the speaker's intended meaning.
39
+ **New Vocabulary:** Change vocabulary to be on Piece-level rather than character-level with vocabulary size 500 piece.
40
 
41
 
42
  ## Usage
 
99
  |Paramter|Value|
100
  |-|-|
101
  |per_device_train_batch_size|6|
102
+ |per_device_eval_batch_size|16|
103
+ |gradient_accumulation_steps|12|
104
  |eval_accumulation_steps|16|
105
  |dataloader_num_workers|2|
106
+ |learning_rate|5e-5|
107
  |adafactor|True|
108
+ |weight_decay|0.08989525|
109
+ |max_grad_norm|0.58585|
110
+ |num_train_epochs|5|
111
+ |warmup_ratio|0.7|
112
  |lr_scheduler_type|constant_with_warmup|
113
  |fp16|True|
114
  |gradient_checkpointing|True|
115
  |sortish_sampler|True|
116
 
117
  ##### Results
118
+ **Train loss:** 0.8925
119
  |Split|Word Error Rate (%)|
120
  |-|-|
121
+ |dev|44.8|
122
+ |tst-HE|39.1|
123
+ |tst-COMMON|43.2|
124
 
125
 
126
  ## Citation