Simonom commited on
Commit
da46df5
·
verified ·
1 Parent(s): 81a8915

Simonom/uzbek-stt-v2-3k-data

Browse files
README.md CHANGED
@@ -9,8 +9,6 @@ datasets:
9
  model-index:
10
  - name: temp
11
  results: []
12
- language:
13
- - uz
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -20,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the common_voice_14_0 dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.6938
24
 
25
  ## Model description
26
 
@@ -40,8 +38,8 @@ More information needed
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 0.001
43
- - train_batch_size: 32
44
- - eval_batch_size: 32
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
@@ -53,8 +51,8 @@ The following hyperparameters were used during training:
53
 
54
  | Training Loss | Epoch | Step | Validation Loss |
55
  |:-------------:|:-----:|:----:|:---------------:|
56
- | 0.8861 | 1.0 | 189 | 0.7931 |
57
- | 0.5809 | 2.0 | 378 | 0.6938 |
58
 
59
 
60
  ### Framework versions
@@ -62,5 +60,5 @@ The following hyperparameters were used during training:
62
  - PEFT 0.11.2.dev0
63
  - Transformers 4.42.0.dev0
64
  - Pytorch 2.3.0+cu121
65
- - Datasets 2.19.1
66
  - Tokenizers 0.19.1
 
9
  model-index:
10
  - name: temp
11
  results: []
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
18
 
19
  This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the common_voice_14_0 dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.8801
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 0.001
41
+ - train_batch_size: 2
42
+ - eval_batch_size: 2
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
+ | 1.1475 | 1.0 | 604 | 1.0288 |
55
+ | 0.6366 | 2.0 | 1208 | 0.8801 |
56
 
57
 
58
  ### Framework versions
 
60
  - PEFT 0.11.2.dev0
61
  - Transformers 4.42.0.dev0
62
  - Pytorch 2.3.0+cu121
63
+ - Datasets 2.19.2
64
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
- "base_model_class": "PeftModel",
5
- "parent_library": "peft.peft_model"
6
  },
7
- "base_model_name_or_path": null,
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
@@ -13,13 +13,13 @@
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 64,
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
- "r": 32,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
+ "base_model_class": "WhisperForConditionalGeneration",
5
+ "parent_library": "transformers.models.whisper.modeling_whisper"
6
  },
7
+ "base_model_name_or_path": "openai/whisper-large-v2",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
 
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
+ "lora_alpha": 16,
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
+ "r": 8,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47d6f931bab68e449a5f041b7b521fe6147eba1f59b2d3bdaef1ba2fffe180f0
3
- size 62976168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fa321b89f4be63edef21522512d29aed96195d8313a5904f591e375d3f37d11
3
+ size 15782928
runs/Jun05_07-54-05_9e0960f27b2e/events.out.tfevents.1717574046.9e0960f27b2e.255.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b149120b90f8b045c9913cca28fb5e5686cafef681282b503d3ab1f2c75e1b1b
3
+ size 4184
runs/Jun05_07-59-36_9e0960f27b2e/events.out.tfevents.1717574377.9e0960f27b2e.5050.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bf8f5cca55114554969c73a3e92fc8e726a48eafe10e71737dc62e14ca62f30
3
+ size 17134
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe94d8a6fe9a79bc68aa0ff884729f1068cb3d5495d760d00e016f711b1a0155
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd8f2faebce869f7c0ec1cbd287a5ff1fa267edd505d8da47488e2106401f87a
3
  size 5240