ggmbr commited on
Commit
31ceba2
·
1 Parent(s): 35a800e

Upload model

Browse files
Files changed (1) hide show
  1. config.json +4 -4
config.json CHANGED
@@ -6,7 +6,7 @@
6
  "add_adapter": false,
7
  "apply_spec_augment": true,
8
  "architectures": [
9
- "WavLMModel"
10
  ],
11
  "attention_dropout": 0.1,
12
  "bos_token_id": 1,
@@ -45,6 +45,7 @@
45
  "ctc_zero_infinity": false,
46
  "diversity_loss_weight": 0.1,
47
  "do_stable_layer_norm": true,
 
48
  "eos_token_id": 2,
49
  "feat_extract_activation": "gelu",
50
  "feat_extract_dropout": 0.0,
@@ -57,6 +58,7 @@
57
  "hidden_dropout": 0.1,
58
  "hidden_size": 1024,
59
  "initializer_range": 0.02,
 
60
  "intermediate_size": 4096,
61
  "layer_norm_eps": 1e-05,
62
  "layerdrop": 0.1,
@@ -117,7 +119,5 @@
117
  "transformers_version": "4.48.2",
118
  "use_weighted_layer_sum": false,
119
  "vocab_size": 32,
120
- "xvector_output_dim": 512,
121
- "embd_size": 250,
122
- "interm_size": 512
123
  }
 
6
  "add_adapter": false,
7
  "apply_spec_augment": true,
8
  "architectures": [
9
+ "NTModel"
10
  ],
11
  "attention_dropout": 0.1,
12
  "bos_token_id": 1,
 
45
  "ctc_zero_infinity": false,
46
  "diversity_loss_weight": 0.1,
47
  "do_stable_layer_norm": true,
48
+ "embd_size": 250,
49
  "eos_token_id": 2,
50
  "feat_extract_activation": "gelu",
51
  "feat_extract_dropout": 0.0,
 
58
  "hidden_dropout": 0.1,
59
  "hidden_size": 1024,
60
  "initializer_range": 0.02,
61
+ "interm_size": 512,
62
  "intermediate_size": 4096,
63
  "layer_norm_eps": 1e-05,
64
  "layerdrop": 0.1,
 
119
  "transformers_version": "4.48.2",
120
  "use_weighted_layer_sum": false,
121
  "vocab_size": 32,
122
+ "xvector_output_dim": 512
 
 
123
  }