fitlemon commited on
Commit
4740922
·
verified ·
1 Parent(s): 9030bb4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -4
README.md CHANGED
@@ -9,6 +9,13 @@ metrics:
9
  model-index:
10
  - name: whisper-small-uz-en-ru-lang-id
11
  results: []
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -16,7 +23,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # whisper-small-uz-en-ru-lang-id
18
 
19
- This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
  - Loss: 0.2065
22
  - Accuracy: 0.9747
@@ -31,10 +38,24 @@ More information needed
31
  More information needed
32
 
33
  ## Training and evaluation data
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
- More information needed
36
 
37
- ## Training procedure
38
 
39
  ### Training hyperparameters
40
 
@@ -65,4 +86,4 @@ The following hyperparameters were used during training:
65
  - Transformers 4.38.2
66
  - Pytorch 2.2.1+cu121
67
  - Datasets 2.17.1
68
- - Tokenizers 0.15.2
 
9
  model-index:
10
  - name: whisper-small-uz-en-ru-lang-id
11
  results: []
12
+ datasets:
13
+ - mozilla-foundation/common_voice_16_1
14
+ language:
15
+ - uz
16
+ - en
17
+ - ru
18
+ pipeline_tag: audio-classification
19
  ---
20
 
21
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
23
 
24
  # whisper-small-uz-en-ru-lang-id
25
 
26
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the "mozilla-foundation/common_voice_16_1"(uz/en/ru) dataset.
27
  It achieves the following results on the evaluation set:
28
  - Loss: 0.2065
29
  - Accuracy: 0.9747
 
38
  More information needed
39
 
40
  ## Training and evaluation data
41
+ ```
42
+ # datasets for each lang-id
43
+ common_voice_train_uz = load_dataset("mozilla-foundation/common_voice_16_1", "uz", split='train', trust_remote_code=True, token=env('HUGGING_TOKEN'), streaming=True)
44
+ common_voice_train_ru = load_dataset("mozilla-foundation/common_voice_16_1", "ru", split='train', trust_remote_code=True, token=env('HUGGING_TOKEN'), streaming=True)
45
+ common_voice_train_en = load_dataset("mozilla-foundation/common_voice_16_1", "en", split='train', trust_remote_code=True, token=env('HUGGING_TOKEN'), streaming=True)
46
+ common_voice_valid_uz = load_dataset("mozilla-foundation/common_voice_16_1", "uz", split='validation', trust_remote_code=True, token=env('HUGGING_TOKEN'), streaming=True)
47
+ common_voice_valid_ru = load_dataset("mozilla-foundation/common_voice_16_1", "ru", split='validation', trust_remote_code=True, token=env('HUGGING_TOKEN'), streaming=True)
48
+ common_voice_valid_en = load_dataset("mozilla-foundation/common_voice_16_1", "en", split='validation', trust_remote_code=True, token=env('HUGGING_TOKEN'), streaming=True)
49
+
50
+ # code to shuffle and to take limited size of data
51
+ ...
52
+ # concatenate 3 datasets
53
+ common_voice['train'] = concatenate_datasets([common_voice_train_uz, common_voice_train_ru, common_voice_train_en])
54
+ ```
55
+ ## Training procedure
56
 
57
+ Used Trainer from transformers
58
 
 
59
 
60
  ### Training hyperparameters
61
 
 
86
  - Transformers 4.38.2
87
  - Pytorch 2.2.1+cu121
88
  - Datasets 2.17.1
89
+ - Tokenizers 0.15.2