devkyle commited on
Commit
7b7bad1
·
verified ·
1 Parent(s): 4504933

End of training

Browse files
README.md CHANGED
@@ -4,8 +4,6 @@ license: apache-2.0
4
  base_model: openai/whisper-base
5
  tags:
6
  - generated_from_trainer
7
- metrics:
8
- - wer
9
  model-index:
10
  - name: whisper-base-v4
11
  results: []
@@ -18,8 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.1814
22
- - Wer: 43.8709
 
 
 
 
 
23
 
24
  ## Model description
25
 
@@ -38,29 +41,19 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 0.0002
42
  - train_batch_size: 16
43
  - eval_batch_size: 8
44
  - seed: 42
45
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 500
48
  - training_steps: 2000
49
  - mixed_precision_training: Native AMP
50
 
51
- ### Training results
52
-
53
- | Training Loss | Epoch | Step | Validation Loss | Wer |
54
- |:-------------:|:-----:|:----:|:---------------:|:-------:|
55
- | 0.4275 | 10.0 | 500 | 0.8866 | 50.2836 |
56
- | 0.069 | 20.0 | 1000 | 1.0774 | 47.0699 |
57
- | 0.0134 | 30.0 | 1500 | 1.1680 | 44.5689 |
58
- | 0.002 | 40.0 | 2000 | 1.1814 | 43.8709 |
59
-
60
-
61
  ### Framework versions
62
 
63
- - Transformers 4.45.2
64
  - Pytorch 2.5.0+cu121
65
- - Datasets 3.0.2
66
- - Tokenizers 0.20.1
 
4
  base_model: openai/whisper-base
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: whisper-base-v4
9
  results: []
 
16
 
17
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - eval_loss: 2.6242
20
+ - eval_wer: 90.4248
21
+ - eval_runtime: 99.7998
22
+ - eval_samples_per_second: 2.004
23
+ - eval_steps_per_second: 0.251
24
+ - epoch: 20.0
25
+ - step: 1000
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 0.0001
45
  - train_batch_size: 16
46
  - eval_batch_size: 8
47
  - seed: 42
48
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
  - lr_scheduler_type: linear
50
  - lr_scheduler_warmup_steps: 500
51
  - training_steps: 2000
52
  - mixed_precision_training: Native AMP
53
 
 
 
 
 
 
 
 
 
 
 
54
  ### Framework versions
55
 
56
+ - Transformers 4.46.2
57
  - Pytorch 2.5.0+cu121
58
+ - Datasets 3.1.0
59
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -1,36 +1,44 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 3,
5
- 1
6
  ],
7
  [
8
- 4,
9
- 2
10
  ],
11
  [
12
- 4,
13
- 3
 
 
 
 
14
  ],
15
  [
16
- 4,
17
  7
18
  ],
19
  [
20
- 5,
21
- 1
22
  ],
23
  [
24
- 5,
25
- 2
26
  ],
27
  [
28
- 5,
29
- 4
30
  ],
31
  [
32
- 5,
33
- 6
 
 
 
 
34
  ]
35
  ],
36
  "begin_suppress_tokens": [
@@ -39,8 +47,17 @@
39
  ],
40
  "bos_token_id": 50257,
41
  "decoder_start_token_id": 50258,
42
- "dropout": 0.1,
43
  "eos_token_id": 50257,
 
 
 
 
 
 
 
 
 
 
44
  "is_multilingual": true,
45
  "lang_to_id": {
46
  "<|af|>": 50327,
@@ -244,5 +261,5 @@
244
  "transcribe": 50359,
245
  "translate": 50358
246
  },
247
- "transformers_version": "4.45.2"
248
  }
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 5,
5
+ 3
6
  ],
7
  [
8
+ 5,
9
+ 9
10
  ],
11
  [
12
+ 8,
13
+ 0
14
+ ],
15
+ [
16
+ 8,
17
+ 4
18
  ],
19
  [
20
+ 8,
21
  7
22
  ],
23
  [
24
+ 8,
25
+ 8
26
  ],
27
  [
28
+ 9,
29
+ 0
30
  ],
31
  [
32
+ 9,
33
+ 7
34
  ],
35
  [
36
+ 9,
37
+ 9
38
+ ],
39
+ [
40
+ 10,
41
+ 5
42
  ]
43
  ],
44
  "begin_suppress_tokens": [
 
47
  ],
48
  "bos_token_id": 50257,
49
  "decoder_start_token_id": 50258,
 
50
  "eos_token_id": 50257,
51
+ "forced_decoder_ids": [
52
+ [
53
+ 1,
54
+ null
55
+ ],
56
+ [
57
+ 2,
58
+ 50359
59
+ ]
60
+ ],
61
  "is_multilingual": true,
62
  "lang_to_id": {
63
  "<|af|>": 50327,
 
261
  "transcribe": 50359,
262
  "translate": 50358
263
  },
264
+ "transformers_version": "4.46.2"
265
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38f3ca61eaf6db16f6996e30b8d2af7af4c00cfe5e55a8fd2bce8deb2c551589
3
  size 437763864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a316a06183fc77f742287a5ef65925c935d70fbf40bee841991c389eb21e5c58
3
  size 437763864
runs/Nov07_03-43-44_9817841b8cc1/events.out.tfevents.1730951035.9817841b8cc1.753.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74bf171e804ba8edbcc67b2c15ef06a782b2c78468d3f38c690516e2f321bf6f
3
- size 15091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea1f4cff0ba2935bfccd063e9edb302bcae18c169481ad432a9e787b0e23b0c1
3
+ size 15935