azizbekphd commited on
Commit
0ec5ef4
·
verified ·
1 Parent(s): 8724c14

Training in progress, epoch 1

Browse files
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "</s>": 27,
3
+ "<s>": 26
4
+ }
config.json CHANGED
@@ -43,7 +43,7 @@
43
  2
44
  ],
45
  "ctc_loss_reduction": "sum",
46
- "ctc_zero_infinity": true,
47
  "diversity_loss_weight": 0.1,
48
  "do_stable_layer_norm": false,
49
  "eos_token_id": 2,
@@ -105,6 +105,6 @@
105
  "torch_dtype": "float32",
106
  "transformers_version": "4.46.2",
107
  "use_weighted_layer_sum": false,
108
- "vocab_size": 26,
109
  "xvector_output_dim": 512
110
  }
 
43
  2
44
  ],
45
  "ctc_loss_reduction": "sum",
46
+ "ctc_zero_infinity": false,
47
  "diversity_loss_weight": 0.1,
48
  "do_stable_layer_norm": false,
49
  "eos_token_id": 2,
 
105
  "torch_dtype": "float32",
106
  "transformers_version": "4.46.2",
107
  "use_weighted_layer_sum": false,
108
+ "vocab_size": 28,
109
  "xvector_output_dim": 512
110
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31f62f6e581e193174f618df93bd1dfd45840e9129651e3560691a29e9a21f2c
3
- size 377592664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d19cf9174fe11f913188adc314eccbaf942340e10c982f64f1de13f76423445
3
+ size 377598816
runs/Nov13_06-41-50_64e20631466d/events.out.tfevents.1731480114.64e20631466d.1557.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5dcc88c8226ce71971d9ffd6a9ffb7a68baae5bcd947d933bbef33a31c9d046
3
+ size 6801
runs/Nov13_07-01-14_64e20631466d/events.out.tfevents.1731481277.64e20631466d.1557.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87137c35d187dbfd09990bfd688821c762e340ec7d1b42d46e6198f2e5b1a29a
3
+ size 6545
runs/Nov13_07-05-48_64e20631466d/events.out.tfevents.1731481549.64e20631466d.1557.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c7ceec6133dd8433c6f2a03d1cdfc249c4e886384238b4a7335f61b90407632
3
+ size 13002
runs/Nov13_07-06-12_64e20631466d/events.out.tfevents.1731481576.64e20631466d.1557.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:175e162348496218f36e614207ca4570c2b76bae2752b092c5593efd81e6ff79
3
+ size 8526
tokenizer_config.json CHANGED
@@ -1,48 +1,46 @@
1
  {
2
  "added_tokens_decoder": {
3
- "0": {
4
- "content": "<pad>",
5
  "lstrip": true,
6
  "normalized": false,
7
  "rstrip": true,
8
  "single_word": false,
9
  "special": false
10
  },
11
- "1": {
12
- "content": "<s>",
13
  "lstrip": true,
14
  "normalized": false,
15
  "rstrip": true,
16
  "single_word": false,
17
  "special": false
18
  },
19
- "2": {
20
- "content": "</s>",
21
- "lstrip": true,
22
  "normalized": false,
23
- "rstrip": true,
24
  "single_word": false,
25
- "special": false
26
  },
27
- "3": {
28
- "content": "<unk>",
29
- "lstrip": true,
30
  "normalized": false,
31
- "rstrip": true,
32
  "single_word": false,
33
- "special": false
34
  }
35
  },
36
  "bos_token": "<s>",
37
  "clean_up_tokenization_spaces": false,
38
  "do_lower_case": false,
39
- "do_normalize": true,
40
  "eos_token": "</s>",
41
  "model_max_length": 1000000000000000019884624838656,
42
  "pad_token": "<pad>",
43
  "processor_class": "Wav2Vec2Processor",
44
  "replace_word_delimiter_char": " ",
45
- "return_attention_mask": false,
46
  "target_lang": null,
47
  "tokenizer_class": "Wav2Vec2CTCTokenizer",
48
  "unk_token": "<unk>",
 
1
  {
2
  "added_tokens_decoder": {
3
+ "24": {
4
+ "content": "<unk>",
5
  "lstrip": true,
6
  "normalized": false,
7
  "rstrip": true,
8
  "single_word": false,
9
  "special": false
10
  },
11
+ "25": {
12
+ "content": "<pad>",
13
  "lstrip": true,
14
  "normalized": false,
15
  "rstrip": true,
16
  "single_word": false,
17
  "special": false
18
  },
19
+ "26": {
20
+ "content": "<s>",
21
+ "lstrip": false,
22
  "normalized": false,
23
+ "rstrip": false,
24
  "single_word": false,
25
+ "special": true
26
  },
27
+ "27": {
28
+ "content": "</s>",
29
+ "lstrip": false,
30
  "normalized": false,
31
+ "rstrip": false,
32
  "single_word": false,
33
+ "special": true
34
  }
35
  },
36
  "bos_token": "<s>",
37
  "clean_up_tokenization_spaces": false,
38
  "do_lower_case": false,
 
39
  "eos_token": "</s>",
40
  "model_max_length": 1000000000000000019884624838656,
41
  "pad_token": "<pad>",
42
  "processor_class": "Wav2Vec2Processor",
43
  "replace_word_delimiter_char": " ",
 
44
  "target_lang": null,
45
  "tokenizer_class": "Wav2Vec2CTCTokenizer",
46
  "unk_token": "<unk>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:410e3ff51388955dfcb2efafe0f0d68666e38092eb6a50c31b708e33080be587
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4b381511a8c2e753d1038b1d2a192792a979f24a813296c8c44cd936355726a
3
  size 5304
vocab.json CHANGED
@@ -1,34 +1,28 @@
1
  {
2
- "'": 27,
3
- "</s>": 2,
4
- "<pad>": 0,
5
- "<s>": 1,
6
- "<unk>": 3,
7
- "A": 7,
8
- "B": 24,
9
- "C": 19,
10
- "D": 14,
11
- "E": 5,
12
- "F": 20,
13
- "G": 21,
14
- "H": 11,
15
- "I": 10,
16
- "J": 29,
17
- "K": 26,
18
- "L": 15,
19
- "M": 17,
20
- "N": 9,
21
- "O": 8,
22
- "P": 23,
23
- "Q": 30,
24
- "R": 13,
25
- "S": 12,
26
- "T": 6,
27
- "U": 16,
28
- "V": 25,
29
- "W": 18,
30
- "X": 28,
31
- "Y": 22,
32
- "Z": 31,
33
- "|": 4
34
  }
 
1
  {
2
+ "<pad>": 25,
3
+ "<unk>": 24,
4
+ "ح": 8,
5
+ "ص": 10,
6
+ "ف": 16,
7
+ "ق": 0,
8
+ "ك": 15,
9
+ "ل": 2,
10
+ "م": 11,
11
+ "ن": 17,
12
+ "و": 4,
13
+ "ي": 12,
14
+ "ٌ": 14,
15
+ "ِ": 13,
16
+ "ٞ": 7,
17
+ "ڶ": 6,
18
+ "ڽ": 22,
19
+ "ھ": 3,
20
+ "ۮ": 9,
21
+ "ݚ": 18,
22
+ "": 21,
23
+ "": 20,
24
+ "": 23,
25
+ "": 19,
26
+ "": 5,
27
+ "": 1
 
 
 
 
 
 
28
  }