gvaya-bsc commited on
Commit
4347a5a
·
verified ·
1 Parent(s): b5d556e

Reupload with correct weights

Browse files
Files changed (2) hide show
  1. config.json +13 -209
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,232 +1,36 @@
1
  {
2
  "_name_or_path": "/gpfs/projects/bsc14/gvaya/GLiClass/model_final",
3
- "architecture_type": "uni-encoder",
4
  "architectures": [
5
- "DebertaForSequenceClassification"
6
  ],
7
  "attention_probs_dropout_prob": 0.1,
8
- "class_token_index": 128001,
9
- "contrastive_loss_coef": 0.0,
10
- "embed_class_token": true,
11
- "encoder_config": {
12
- "_attn_implementation_autoset": false,
13
- "_name_or_path": "./deberta",
14
- "add_cross_attention": false,
15
- "architectures": [
16
- "DebertaV2Model"
17
- ],
18
- "attention_probs_dropout_prob": 0.1,
19
- "bad_words_ids": null,
20
- "begin_suppress_tokens": null,
21
- "bos_token_id": null,
22
- "chunk_size_feed_forward": 0,
23
- "cross_attention_hidden_size": null,
24
- "decoder_start_token_id": null,
25
- "diversity_penalty": 0.0,
26
- "do_sample": false,
27
- "early_stopping": false,
28
- "encoder_no_repeat_ngram_size": 0,
29
- "eos_token_id": null,
30
- "exponential_decay_length_penalty": null,
31
- "finetuning_task": null,
32
- "forced_bos_token_id": null,
33
- "forced_eos_token_id": null,
34
- "hidden_act": "gelu",
35
- "hidden_dropout_prob": 0.1,
36
- "hidden_size": 768,
37
- "id2label": {
38
- "0": "LABEL_0",
39
- "1": "LABEL_1"
40
- },
41
- "initializer_range": 0.02,
42
- "intermediate_size": 3072,
43
- "is_decoder": false,
44
- "is_encoder_decoder": false,
45
- "label2id": {
46
- "LABEL_0": 0,
47
- "LABEL_1": 1
48
- },
49
- "layer_norm_eps": 1e-07,
50
- "legacy": true,
51
- "length_penalty": 1.0,
52
- "max_length": 20,
53
- "max_position_embeddings": 512,
54
- "max_relative_positions": -1,
55
- "min_length": 0,
56
- "model_type": "deberta-v2",
57
- "no_repeat_ngram_size": 0,
58
- "norm_rel_ebd": "layer_norm",
59
- "num_attention_heads": 12,
60
- "num_beam_groups": 1,
61
- "num_beams": 1,
62
- "num_hidden_layers": 6,
63
- "num_return_sequences": 1,
64
- "output_attentions": false,
65
- "output_hidden_states": false,
66
- "output_scores": false,
67
- "pad_token_id": 0,
68
- "pooler_dropout": 0,
69
- "pooler_hidden_act": "gelu",
70
- "pooler_hidden_size": 768,
71
- "pos_att_type": [
72
- "p2c",
73
- "c2p"
74
- ],
75
- "position_biased_input": false,
76
- "position_buckets": 256,
77
- "prefix": null,
78
- "problem_type": null,
79
- "pruned_heads": {},
80
- "relative_attention": true,
81
- "remove_invalid_values": false,
82
- "repetition_penalty": 1.0,
83
- "return_dict": true,
84
- "return_dict_in_generate": false,
85
- "sep_token_id": null,
86
- "share_att_key": true,
87
- "suppress_tokens": null,
88
- "task_specific_params": null,
89
- "temperature": 1.0,
90
- "tf_legacy_loss": false,
91
- "tie_encoder_decoder": false,
92
- "tie_word_embeddings": true,
93
- "tokenizer_class": null,
94
- "top_k": 50,
95
- "top_p": 1.0,
96
- "torch_dtype": "float32",
97
- "torchscript": false,
98
- "type_vocab_size": 0,
99
- "typical_p": 1.0,
100
- "use_bfloat16": false,
101
- "vocab_size": 128003
102
- },
103
- "encoder_model_name": "./deberta",
104
- "extract_text_features": false,
105
- "focal_loss_alpha": -1,
106
- "focal_loss_gamma": -1,
107
  "hidden_act": "gelu",
108
  "hidden_dropout_prob": 0.1,
109
  "hidden_size": 768,
110
- "id2label": {
111
- "0": "LABEL_0",
112
- "1": "LABEL_1",
113
- "2": "LABEL_2"
114
- },
115
- "ignore_index": -100,
116
- "initializer_range": 0.03,
117
  "intermediate_size": 3072,
118
- "label2id": {
119
- "LABEL_0": 0,
120
- "LABEL_1": 1,
121
- "LABEL_2": 2
122
- },
123
- "label_model_config": {
124
- "_attn_implementation_autoset": false,
125
- "_name_or_path": "BAAI/bge-small-en-v1.5",
126
- "add_cross_attention": false,
127
- "architectures": [
128
- "BertModel"
129
- ],
130
- "attention_probs_dropout_prob": 0.1,
131
- "bad_words_ids": null,
132
- "begin_suppress_tokens": null,
133
- "bos_token_id": null,
134
- "chunk_size_feed_forward": 0,
135
- "classifier_dropout": null,
136
- "cross_attention_hidden_size": null,
137
- "decoder_start_token_id": null,
138
- "diversity_penalty": 0.0,
139
- "do_sample": false,
140
- "early_stopping": false,
141
- "encoder_no_repeat_ngram_size": 0,
142
- "eos_token_id": null,
143
- "exponential_decay_length_penalty": null,
144
- "finetuning_task": null,
145
- "forced_bos_token_id": null,
146
- "forced_eos_token_id": null,
147
- "hidden_act": "gelu",
148
- "hidden_dropout_prob": 0.1,
149
- "hidden_size": 384,
150
- "id2label": {
151
- "0": "LABEL_0"
152
- },
153
- "initializer_range": 0.02,
154
- "intermediate_size": 1536,
155
- "is_decoder": false,
156
- "is_encoder_decoder": false,
157
- "label2id": {
158
- "LABEL_0": 0
159
- },
160
- "layer_norm_eps": 1e-12,
161
- "length_penalty": 1.0,
162
- "max_length": 20,
163
- "max_position_embeddings": 512,
164
- "min_length": 0,
165
- "model_type": "bert",
166
- "no_repeat_ngram_size": 0,
167
- "num_attention_heads": 12,
168
- "num_beam_groups": 1,
169
- "num_beams": 1,
170
- "num_hidden_layers": 12,
171
- "num_return_sequences": 1,
172
- "output_attentions": false,
173
- "output_hidden_states": false,
174
- "output_scores": false,
175
- "pad_token_id": 0,
176
- "position_embedding_type": "absolute",
177
- "prefix": null,
178
- "problem_type": null,
179
- "pruned_heads": {},
180
- "remove_invalid_values": false,
181
- "repetition_penalty": 1.0,
182
- "return_dict": true,
183
- "return_dict_in_generate": false,
184
- "sep_token_id": null,
185
- "suppress_tokens": null,
186
- "task_specific_params": null,
187
- "temperature": 1.0,
188
- "tf_legacy_loss": false,
189
- "tie_encoder_decoder": false,
190
- "tie_word_embeddings": true,
191
- "tokenizer_class": null,
192
- "top_k": 50,
193
- "top_p": 1.0,
194
- "torch_dtype": "float32",
195
- "torchscript": false,
196
- "type_vocab_size": 2,
197
- "typical_p": 1.0,
198
- "use_bfloat16": false,
199
- "use_cache": true,
200
- "vocab_size": 30522
201
- },
202
- "label_model_name": "BAAI/bge-small-en-v1.5",
203
  "layer_norm_eps": 1e-07,
204
  "legacy": true,
205
- "logit_scale_init_value": 2.6592,
206
- "max_num_classes": 25,
207
  "max_position_embeddings": 512,
208
  "max_relative_positions": -1,
209
- "model_type": "deberta",
210
- "normalize_features": false,
211
  "num_attention_heads": 12,
212
- "num_hidden_layers": 12,
213
  "pad_token_id": 0,
214
  "pooler_dropout": 0,
215
  "pooler_hidden_act": "gelu",
216
  "pooler_hidden_size": 768,
217
- "pooling_strategy": "avg",
218
- "pos_att_type": null,
219
- "position_biased_input": true,
220
- "problem_type": "single_label_classification",
221
- "projector_hidden_act": "gelu",
222
- "prompt_first": true,
223
- "relative_attention": false,
224
- "scorer_type": "simple",
225
- "squeeze_layers": false,
226
- "text_token_index": 128002,
227
  "torch_dtype": "float32",
228
  "transformers_version": "4.48.2",
229
  "type_vocab_size": 0,
230
- "use_lstm": false,
231
  "vocab_size": 128003
232
  }
 
1
  {
2
  "_name_or_path": "/gpfs/projects/bsc14/gvaya/GLiClass/model_final",
 
3
  "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
10
+ "initializer_range": 0.02,
 
 
 
 
 
 
11
  "intermediate_size": 3072,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  "layer_norm_eps": 1e-07,
13
  "legacy": true,
 
 
14
  "max_position_embeddings": 512,
15
  "max_relative_positions": -1,
16
+ "model_type": "deberta-v2",
17
+ "norm_rel_ebd": "layer_norm",
18
  "num_attention_heads": 12,
19
+ "num_hidden_layers": 6,
20
  "pad_token_id": 0,
21
  "pooler_dropout": 0,
22
  "pooler_hidden_act": "gelu",
23
  "pooler_hidden_size": 768,
24
+ "pos_att_type": [
25
+ "p2c",
26
+ "c2p"
27
+ ],
28
+ "position_biased_input": false,
29
+ "position_buckets": 256,
30
+ "relative_attention": true,
31
+ "share_att_key": true,
 
 
32
  "torch_dtype": "float32",
33
  "transformers_version": "4.48.2",
34
  "type_vocab_size": 0,
 
35
  "vocab_size": 128003
36
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3cd2e61f23063a5d0df96dbb92b4b19f5ae8417047b01adbcfbdb6bd719f70bc
3
- size 737376356
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed30846297ffbfda9dafbcf158af88e5528e40486a7f10fab946b6ed663602fe
3
+ size 567300568