luukschmitz commited on
Commit
5ce3df9
·
verified ·
1 Parent(s): 8557eff

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - text-classification
6
+ base_model: microsoft/deberta-v3-base
7
+ widget:
8
+ - text: "I love AutoTrain"
9
+ ---
10
+
11
+ # Model Trained Using AutoTrain
12
+
13
+ - Problem type: Text Classification
14
+
15
+ ## Validation Metrics
16
+ loss: 0.9531590342521667
17
+
18
+ f1_macro: 0.4940971741477817
19
+
20
+ f1_micro: 0.6548117154811716
21
+
22
+ f1_weighted: 0.628735443682528
23
+
24
+ precision_macro: 0.6898690717515761
25
+
26
+ precision_micro: 0.6548117154811716
27
+
28
+ precision_weighted: 0.6835107777099136
29
+
30
+ recall_macro: 0.47932568993970953
31
+
32
+ recall_micro: 0.6548117154811716
33
+
34
+ recall_weighted: 0.6548117154811716
35
+
36
+ accuracy: 0.6548117154811716
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
checkpoint-126/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-base",
3
+ "_num_labels": 4,
4
+ "architectures": [
5
+ "DebertaV2ForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "0",
13
+ "1": "1",
14
+ "2": "2",
15
+ "3": "3"
16
+ },
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "0": 0,
21
+ "1": 1,
22
+ "2": 2,
23
+ "3": 3
24
+ },
25
+ "layer_norm_eps": 1e-07,
26
+ "max_position_embeddings": 512,
27
+ "max_relative_positions": -1,
28
+ "model_type": "deberta-v2",
29
+ "norm_rel_ebd": "layer_norm",
30
+ "num_attention_heads": 12,
31
+ "num_hidden_layers": 12,
32
+ "pad_token_id": 0,
33
+ "pooler_dropout": 0,
34
+ "pooler_hidden_act": "gelu",
35
+ "pooler_hidden_size": 768,
36
+ "pos_att_type": [
37
+ "p2c",
38
+ "c2p"
39
+ ],
40
+ "position_biased_input": false,
41
+ "position_buckets": 256,
42
+ "relative_attention": true,
43
+ "share_att_key": true,
44
+ "torch_dtype": "float32",
45
+ "transformers_version": "4.42.2",
46
+ "type_vocab_size": 0,
47
+ "vocab_size": 128100
48
+ }
checkpoint-126/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b455d855ad607c9d001e2714b45efe7ac989414765a467165601716df6bb515c
3
+ size 737725432
checkpoint-126/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e067d89323a1a4700dac7fea00b51ac01178c2b3e0801d723694baa9eb06a71
3
+ size 1475570682
checkpoint-126/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e34166419e58435289b4f8cde27e4dbd2b3646320aa60d17bb4c1144678e1b8
3
+ size 14244
checkpoint-126/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9c966dc6233e3ceeec2c56e1c1e169ae5eccb49c9d5b67c96059961ed9d89ce
3
+ size 1064
checkpoint-126/trainer_state.json ADDED
@@ -0,0 +1,519 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9531590342521667,
3
+ "best_model_checkpoint": "Geodeberta2/checkpoint-126",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 126,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.031746031746031744,
13
+ "grad_norm": 3.382016897201538,
14
+ "learning_rate": 1.5873015873015873e-06,
15
+ "loss": 1.2687,
16
+ "step": 2
17
+ },
18
+ {
19
+ "epoch": 0.06349206349206349,
20
+ "grad_norm": 3.2359426021575928,
21
+ "learning_rate": 3.1746031746031746e-06,
22
+ "loss": 1.2944,
23
+ "step": 4
24
+ },
25
+ {
26
+ "epoch": 0.09523809523809523,
27
+ "grad_norm": 2.8667497634887695,
28
+ "learning_rate": 4.7619047619047615e-06,
29
+ "loss": 1.2725,
30
+ "step": 6
31
+ },
32
+ {
33
+ "epoch": 0.12698412698412698,
34
+ "grad_norm": 1.7016966342926025,
35
+ "learning_rate": 6.349206349206349e-06,
36
+ "loss": 1.3089,
37
+ "step": 8
38
+ },
39
+ {
40
+ "epoch": 0.15873015873015872,
41
+ "grad_norm": 2.4599409103393555,
42
+ "learning_rate": 7.936507936507936e-06,
43
+ "loss": 1.2902,
44
+ "step": 10
45
+ },
46
+ {
47
+ "epoch": 0.19047619047619047,
48
+ "grad_norm": 2.53178334236145,
49
+ "learning_rate": 9.523809523809523e-06,
50
+ "loss": 1.2766,
51
+ "step": 12
52
+ },
53
+ {
54
+ "epoch": 0.2222222222222222,
55
+ "grad_norm": 3.3443331718444824,
56
+ "learning_rate": 1.1111111111111112e-05,
57
+ "loss": 1.2274,
58
+ "step": 14
59
+ },
60
+ {
61
+ "epoch": 0.25396825396825395,
62
+ "grad_norm": 2.784144878387451,
63
+ "learning_rate": 1.2698412698412699e-05,
64
+ "loss": 1.2141,
65
+ "step": 16
66
+ },
67
+ {
68
+ "epoch": 0.2857142857142857,
69
+ "grad_norm": 3.0724332332611084,
70
+ "learning_rate": 1.4285714285714285e-05,
71
+ "loss": 1.2465,
72
+ "step": 18
73
+ },
74
+ {
75
+ "epoch": 0.31746031746031744,
76
+ "grad_norm": 1.5863182544708252,
77
+ "learning_rate": 1.5873015873015872e-05,
78
+ "loss": 1.2876,
79
+ "step": 20
80
+ },
81
+ {
82
+ "epoch": 0.3492063492063492,
83
+ "grad_norm": 1.5585477352142334,
84
+ "learning_rate": 1.746031746031746e-05,
85
+ "loss": 1.2536,
86
+ "step": 22
87
+ },
88
+ {
89
+ "epoch": 0.38095238095238093,
90
+ "grad_norm": 2.9539477825164795,
91
+ "learning_rate": 1.9047619047619046e-05,
92
+ "loss": 1.1878,
93
+ "step": 24
94
+ },
95
+ {
96
+ "epoch": 0.4126984126984127,
97
+ "grad_norm": 1.7762205600738525,
98
+ "learning_rate": 2.0634920634920636e-05,
99
+ "loss": 1.2631,
100
+ "step": 26
101
+ },
102
+ {
103
+ "epoch": 0.4444444444444444,
104
+ "grad_norm": 3.783928155899048,
105
+ "learning_rate": 2.2222222222222223e-05,
106
+ "loss": 1.0052,
107
+ "step": 28
108
+ },
109
+ {
110
+ "epoch": 0.47619047619047616,
111
+ "grad_norm": 1.4549317359924316,
112
+ "learning_rate": 2.380952380952381e-05,
113
+ "loss": 1.1053,
114
+ "step": 30
115
+ },
116
+ {
117
+ "epoch": 0.5079365079365079,
118
+ "grad_norm": 2.450103521347046,
119
+ "learning_rate": 2.5396825396825397e-05,
120
+ "loss": 1.0221,
121
+ "step": 32
122
+ },
123
+ {
124
+ "epoch": 0.5396825396825397,
125
+ "grad_norm": 3.955317258834839,
126
+ "learning_rate": 2.6984126984126984e-05,
127
+ "loss": 1.1519,
128
+ "step": 34
129
+ },
130
+ {
131
+ "epoch": 0.5714285714285714,
132
+ "grad_norm": 3.0787367820739746,
133
+ "learning_rate": 2.857142857142857e-05,
134
+ "loss": 0.8755,
135
+ "step": 36
136
+ },
137
+ {
138
+ "epoch": 0.6031746031746031,
139
+ "grad_norm": 2.1669228076934814,
140
+ "learning_rate": 3.0158730158730158e-05,
141
+ "loss": 1.0728,
142
+ "step": 38
143
+ },
144
+ {
145
+ "epoch": 0.6349206349206349,
146
+ "grad_norm": 2.1683146953582764,
147
+ "learning_rate": 3.1746031746031745e-05,
148
+ "loss": 1.0817,
149
+ "step": 40
150
+ },
151
+ {
152
+ "epoch": 0.6666666666666666,
153
+ "grad_norm": 2.4933903217315674,
154
+ "learning_rate": 3.3333333333333335e-05,
155
+ "loss": 1.0527,
156
+ "step": 42
157
+ },
158
+ {
159
+ "epoch": 0.6984126984126984,
160
+ "grad_norm": 2.5503716468811035,
161
+ "learning_rate": 3.492063492063492e-05,
162
+ "loss": 1.1201,
163
+ "step": 44
164
+ },
165
+ {
166
+ "epoch": 0.7301587301587301,
167
+ "grad_norm": 2.326674461364746,
168
+ "learning_rate": 3.650793650793651e-05,
169
+ "loss": 0.8091,
170
+ "step": 46
171
+ },
172
+ {
173
+ "epoch": 0.7619047619047619,
174
+ "grad_norm": 2.899217128753662,
175
+ "learning_rate": 3.809523809523809e-05,
176
+ "loss": 0.851,
177
+ "step": 48
178
+ },
179
+ {
180
+ "epoch": 0.7936507936507936,
181
+ "grad_norm": 2.776383399963379,
182
+ "learning_rate": 3.968253968253968e-05,
183
+ "loss": 1.0539,
184
+ "step": 50
185
+ },
186
+ {
187
+ "epoch": 0.8253968253968254,
188
+ "grad_norm": 3.472299814224243,
189
+ "learning_rate": 4.126984126984127e-05,
190
+ "loss": 0.9422,
191
+ "step": 52
192
+ },
193
+ {
194
+ "epoch": 0.8571428571428571,
195
+ "grad_norm": 4.994964122772217,
196
+ "learning_rate": 4.2857142857142856e-05,
197
+ "loss": 1.0585,
198
+ "step": 54
199
+ },
200
+ {
201
+ "epoch": 0.8888888888888888,
202
+ "grad_norm": 7.897761821746826,
203
+ "learning_rate": 4.4444444444444447e-05,
204
+ "loss": 1.1766,
205
+ "step": 56
206
+ },
207
+ {
208
+ "epoch": 0.9206349206349206,
209
+ "grad_norm": 3.434422731399536,
210
+ "learning_rate": 4.603174603174603e-05,
211
+ "loss": 1.0149,
212
+ "step": 58
213
+ },
214
+ {
215
+ "epoch": 0.9523809523809523,
216
+ "grad_norm": 4.816616535186768,
217
+ "learning_rate": 4.761904761904762e-05,
218
+ "loss": 0.8752,
219
+ "step": 60
220
+ },
221
+ {
222
+ "epoch": 0.9841269841269841,
223
+ "grad_norm": 2.6246337890625,
224
+ "learning_rate": 4.9206349206349204e-05,
225
+ "loss": 0.9825,
226
+ "step": 62
227
+ },
228
+ {
229
+ "epoch": 1.0,
230
+ "eval_accuracy": 0.5271966527196653,
231
+ "eval_f1_macro": 0.29448106332528445,
232
+ "eval_f1_micro": 0.5271966527196653,
233
+ "eval_f1_weighted": 0.47887320097744607,
234
+ "eval_loss": 1.0690832138061523,
235
+ "eval_precision_macro": 0.2979542433081334,
236
+ "eval_precision_micro": 0.5271966527196653,
237
+ "eval_precision_weighted": 0.5181255807964261,
238
+ "eval_recall_macro": 0.35484535475707013,
239
+ "eval_recall_micro": 0.5271966527196653,
240
+ "eval_recall_weighted": 0.5271966527196653,
241
+ "eval_runtime": 1.6359,
242
+ "eval_samples_per_second": 292.199,
243
+ "eval_steps_per_second": 4.89,
244
+ "step": 63
245
+ },
246
+ {
247
+ "epoch": 1.0158730158730158,
248
+ "grad_norm": 4.054161071777344,
249
+ "learning_rate": 4.991181657848325e-05,
250
+ "loss": 1.0958,
251
+ "step": 64
252
+ },
253
+ {
254
+ "epoch": 1.0476190476190477,
255
+ "grad_norm": 2.788057565689087,
256
+ "learning_rate": 4.973544973544973e-05,
257
+ "loss": 1.1205,
258
+ "step": 66
259
+ },
260
+ {
261
+ "epoch": 1.0793650793650793,
262
+ "grad_norm": 4.333444595336914,
263
+ "learning_rate": 4.955908289241622e-05,
264
+ "loss": 0.9343,
265
+ "step": 68
266
+ },
267
+ {
268
+ "epoch": 1.1111111111111112,
269
+ "grad_norm": 2.873152732849121,
270
+ "learning_rate": 4.938271604938271e-05,
271
+ "loss": 0.8455,
272
+ "step": 70
273
+ },
274
+ {
275
+ "epoch": 1.1428571428571428,
276
+ "grad_norm": 3.179229736328125,
277
+ "learning_rate": 4.9206349206349204e-05,
278
+ "loss": 0.905,
279
+ "step": 72
280
+ },
281
+ {
282
+ "epoch": 1.1746031746031746,
283
+ "grad_norm": 6.202491760253906,
284
+ "learning_rate": 4.9029982363315695e-05,
285
+ "loss": 0.7701,
286
+ "step": 74
287
+ },
288
+ {
289
+ "epoch": 1.2063492063492063,
290
+ "grad_norm": 7.958104610443115,
291
+ "learning_rate": 4.8853615520282185e-05,
292
+ "loss": 0.9304,
293
+ "step": 76
294
+ },
295
+ {
296
+ "epoch": 1.2380952380952381,
297
+ "grad_norm": 6.333682537078857,
298
+ "learning_rate": 4.8677248677248676e-05,
299
+ "loss": 1.0795,
300
+ "step": 78
301
+ },
302
+ {
303
+ "epoch": 1.2698412698412698,
304
+ "grad_norm": 5.300474166870117,
305
+ "learning_rate": 4.850088183421517e-05,
306
+ "loss": 0.9147,
307
+ "step": 80
308
+ },
309
+ {
310
+ "epoch": 1.3015873015873016,
311
+ "grad_norm": 3.8622822761535645,
312
+ "learning_rate": 4.832451499118166e-05,
313
+ "loss": 0.7233,
314
+ "step": 82
315
+ },
316
+ {
317
+ "epoch": 1.3333333333333333,
318
+ "grad_norm": 3.2566518783569336,
319
+ "learning_rate": 4.814814814814815e-05,
320
+ "loss": 0.9643,
321
+ "step": 84
322
+ },
323
+ {
324
+ "epoch": 1.3650793650793651,
325
+ "grad_norm": 3.669085741043091,
326
+ "learning_rate": 4.797178130511464e-05,
327
+ "loss": 0.5868,
328
+ "step": 86
329
+ },
330
+ {
331
+ "epoch": 1.3968253968253967,
332
+ "grad_norm": 5.4853363037109375,
333
+ "learning_rate": 4.779541446208113e-05,
334
+ "loss": 0.9615,
335
+ "step": 88
336
+ },
337
+ {
338
+ "epoch": 1.4285714285714286,
339
+ "grad_norm": 4.062696933746338,
340
+ "learning_rate": 4.761904761904762e-05,
341
+ "loss": 0.827,
342
+ "step": 90
343
+ },
344
+ {
345
+ "epoch": 1.4603174603174602,
346
+ "grad_norm": 2.2320899963378906,
347
+ "learning_rate": 4.744268077601411e-05,
348
+ "loss": 0.8415,
349
+ "step": 92
350
+ },
351
+ {
352
+ "epoch": 1.492063492063492,
353
+ "grad_norm": 5.46465539932251,
354
+ "learning_rate": 4.72663139329806e-05,
355
+ "loss": 0.7415,
356
+ "step": 94
357
+ },
358
+ {
359
+ "epoch": 1.5238095238095237,
360
+ "grad_norm": 2.736971139907837,
361
+ "learning_rate": 4.708994708994709e-05,
362
+ "loss": 0.8833,
363
+ "step": 96
364
+ },
365
+ {
366
+ "epoch": 1.5555555555555556,
367
+ "grad_norm": 2.601299524307251,
368
+ "learning_rate": 4.691358024691358e-05,
369
+ "loss": 0.7582,
370
+ "step": 98
371
+ },
372
+ {
373
+ "epoch": 1.5873015873015874,
374
+ "grad_norm": 6.022972106933594,
375
+ "learning_rate": 4.673721340388007e-05,
376
+ "loss": 0.7518,
377
+ "step": 100
378
+ },
379
+ {
380
+ "epoch": 1.619047619047619,
381
+ "grad_norm": 6.7905378341674805,
382
+ "learning_rate": 4.656084656084656e-05,
383
+ "loss": 1.0717,
384
+ "step": 102
385
+ },
386
+ {
387
+ "epoch": 1.6507936507936507,
388
+ "grad_norm": 3.2041382789611816,
389
+ "learning_rate": 4.638447971781305e-05,
390
+ "loss": 0.6562,
391
+ "step": 104
392
+ },
393
+ {
394
+ "epoch": 1.6825396825396826,
395
+ "grad_norm": 6.4817681312561035,
396
+ "learning_rate": 4.620811287477954e-05,
397
+ "loss": 0.9219,
398
+ "step": 106
399
+ },
400
+ {
401
+ "epoch": 1.7142857142857144,
402
+ "grad_norm": 4.928455829620361,
403
+ "learning_rate": 4.603174603174603e-05,
404
+ "loss": 0.8101,
405
+ "step": 108
406
+ },
407
+ {
408
+ "epoch": 1.746031746031746,
409
+ "grad_norm": 4.112485885620117,
410
+ "learning_rate": 4.585537918871252e-05,
411
+ "loss": 0.8257,
412
+ "step": 110
413
+ },
414
+ {
415
+ "epoch": 1.7777777777777777,
416
+ "grad_norm": 6.042239189147949,
417
+ "learning_rate": 4.567901234567901e-05,
418
+ "loss": 0.6938,
419
+ "step": 112
420
+ },
421
+ {
422
+ "epoch": 1.8095238095238095,
423
+ "grad_norm": 2.9191532135009766,
424
+ "learning_rate": 4.55026455026455e-05,
425
+ "loss": 0.8586,
426
+ "step": 114
427
+ },
428
+ {
429
+ "epoch": 1.8412698412698414,
430
+ "grad_norm": 2.405008554458618,
431
+ "learning_rate": 4.532627865961199e-05,
432
+ "loss": 0.6218,
433
+ "step": 116
434
+ },
435
+ {
436
+ "epoch": 1.873015873015873,
437
+ "grad_norm": 3.959949016571045,
438
+ "learning_rate": 4.5149911816578484e-05,
439
+ "loss": 0.7355,
440
+ "step": 118
441
+ },
442
+ {
443
+ "epoch": 1.9047619047619047,
444
+ "grad_norm": 3.4744226932525635,
445
+ "learning_rate": 4.4973544973544974e-05,
446
+ "loss": 0.7225,
447
+ "step": 120
448
+ },
449
+ {
450
+ "epoch": 1.9365079365079365,
451
+ "grad_norm": 3.96034836769104,
452
+ "learning_rate": 4.4797178130511465e-05,
453
+ "loss": 0.7276,
454
+ "step": 122
455
+ },
456
+ {
457
+ "epoch": 1.9682539682539684,
458
+ "grad_norm": 4.284174919128418,
459
+ "learning_rate": 4.4620811287477956e-05,
460
+ "loss": 0.6909,
461
+ "step": 124
462
+ },
463
+ {
464
+ "epoch": 2.0,
465
+ "grad_norm": 11.619244575500488,
466
+ "learning_rate": 4.4444444444444447e-05,
467
+ "loss": 0.7797,
468
+ "step": 126
469
+ },
470
+ {
471
+ "epoch": 2.0,
472
+ "eval_accuracy": 0.6548117154811716,
473
+ "eval_f1_macro": 0.4940971741477817,
474
+ "eval_f1_micro": 0.6548117154811716,
475
+ "eval_f1_weighted": 0.628735443682528,
476
+ "eval_loss": 0.9531590342521667,
477
+ "eval_precision_macro": 0.6898690717515761,
478
+ "eval_precision_micro": 0.6548117154811716,
479
+ "eval_precision_weighted": 0.6835107777099136,
480
+ "eval_recall_macro": 0.47932568993970953,
481
+ "eval_recall_micro": 0.6548117154811716,
482
+ "eval_recall_weighted": 0.6548117154811716,
483
+ "eval_runtime": 1.6407,
484
+ "eval_samples_per_second": 291.337,
485
+ "eval_steps_per_second": 4.876,
486
+ "step": 126
487
+ }
488
+ ],
489
+ "logging_steps": 2,
490
+ "max_steps": 630,
491
+ "num_input_tokens_seen": 0,
492
+ "num_train_epochs": 10,
493
+ "save_steps": 500,
494
+ "stateful_callbacks": {
495
+ "EarlyStoppingCallback": {
496
+ "args": {
497
+ "early_stopping_patience": 5,
498
+ "early_stopping_threshold": 0.01
499
+ },
500
+ "attributes": {
501
+ "early_stopping_patience_counter": 0
502
+ }
503
+ },
504
+ "TrainerControl": {
505
+ "args": {
506
+ "should_epoch_stop": false,
507
+ "should_evaluate": false,
508
+ "should_log": false,
509
+ "should_save": true,
510
+ "should_training_stop": false
511
+ },
512
+ "attributes": {}
513
+ }
514
+ },
515
+ "total_flos": 525977876877312.0,
516
+ "train_batch_size": 32,
517
+ "trial_name": null,
518
+ "trial_params": null
519
+ }
checkpoint-126/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fe8a120b203b6004e539160e2facc359a73c35b92c523cb63fc1e9b98bc606d
3
+ size 5176
config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-base",
3
+ "_num_labels": 4,
4
+ "architectures": [
5
+ "DebertaV2ForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "0",
13
+ "1": "1",
14
+ "2": "2",
15
+ "3": "3"
16
+ },
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "0": 0,
21
+ "1": 1,
22
+ "2": 2,
23
+ "3": 3
24
+ },
25
+ "layer_norm_eps": 1e-07,
26
+ "max_position_embeddings": 512,
27
+ "max_relative_positions": -1,
28
+ "model_type": "deberta-v2",
29
+ "norm_rel_ebd": "layer_norm",
30
+ "num_attention_heads": 12,
31
+ "num_hidden_layers": 12,
32
+ "pad_token_id": 0,
33
+ "pooler_dropout": 0,
34
+ "pooler_hidden_act": "gelu",
35
+ "pooler_hidden_size": 768,
36
+ "pos_att_type": [
37
+ "p2c",
38
+ "c2p"
39
+ ],
40
+ "position_biased_input": false,
41
+ "position_buckets": 256,
42
+ "relative_attention": true,
43
+ "share_att_key": true,
44
+ "torch_dtype": "float32",
45
+ "transformers_version": "4.42.2",
46
+ "type_vocab_size": 0,
47
+ "vocab_size": 128100
48
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b455d855ad607c9d001e2714b45efe7ac989414765a467165601716df6bb515c
3
+ size 737725432
runs/Jul13_14-58-37_r-luukschmitz-geogeogeo-33n1b78h-f5b3e-xribo/events.out.tfevents.1720882718.r-luukschmitz-geogeogeo-33n1b78h-f5b3e-xribo.77.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6527b2a99ef4dd24504e8aebe85db2834c83b230894adae0f1346929b4ac8a5
3
- size 5152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c1041a49488cd6e5c8f0c86b1a45d02238d500c176228cf2b58f84232ce4217
3
+ size 57475
runs/Jul13_14-58-37_r-luukschmitz-geogeogeo-33n1b78h-f5b3e-xribo/events.out.tfevents.1720882910.r-luukschmitz-geogeogeo-33n1b78h-f5b3e-xribo.77.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2a1117d24ea39f0d813a41ab1fc00eccd137d8ceab3d05d2c7f9523cde56823
3
+ size 921
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fe8a120b203b6004e539160e2facc359a73c35b92c523cb63fc1e9b98bc606d
3
+ size 5176
training_params.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "Geodeberta2/autotrain-data",
3
+ "model": "microsoft/deberta-v3-base",
4
+ "lr": 5e-05,
5
+ "epochs": 10,
6
+ "max_seq_length": 256,
7
+ "batch_size": 32,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "text_column": "autotrain_text",
18
+ "target_column": "autotrain_label",
19
+ "logging_steps": -1,
20
+ "project_name": "Geodeberta2",
21
+ "auto_find_batch_size": false,
22
+ "mixed_precision": "fp16",
23
+ "save_total_limit": 1,
24
+ "push_to_hub": true,
25
+ "eval_strategy": "epoch",
26
+ "username": "luukschmitz",
27
+ "log": "tensorboard",
28
+ "early_stopping_patience": 5,
29
+ "early_stopping_threshold": 0.01
30
+ }