balhafni commited on
Commit
b86c612
ยท
verified ยท
1 Parent(s): 14099cb

zaebuc nopnx pretrained model

Browse files
config.json ADDED
@@ -0,0 +1,659 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/scratch/ba63/BERT_models/bert-base-arabertv02",
3
+ "architectures": [
4
+ "BertForTokenClassificationSingleLabel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "D*",
13
+ "1": "DDK*",
14
+ "2": "DK*",
15
+ "3": "DKD*",
16
+ "4": "DKI_[\u0644]K*",
17
+ "5": "I_[ ]K*",
18
+ "6": "I_[ ]K*I_[ ]K",
19
+ "7": "I_[ ]K*I_[ ]KI_[ ]K",
20
+ "8": "I_[ ]K*R_[\u0623]",
21
+ "9": "I_[\u0623]K*",
22
+ "10": "I_[\u0623]K*D",
23
+ "11": "I_[\u0623\u0628]KD*",
24
+ "12": "I_[\u0625]K*",
25
+ "13": "I_[\u0625\u0630]KD*",
26
+ "14": "I_[\u0625\u0644]KD*",
27
+ "15": "I_[\u0626]K*",
28
+ "16": "I_[ \u0627]K*",
29
+ "17": "I_[\u0627]K*",
30
+ "18": "I_[\u0627]KDK*",
31
+ "19": "I_[\u0627\u0644]K*",
32
+ "20": "I_[\u0628]K*",
33
+ "21": "I_[\u062a]K*",
34
+ "22": "I_[\u062f]K*",
35
+ "23": "I_[\u0631]K*",
36
+ "24": "I_[\u0639]K*",
37
+ "25": "I_[\u0639]KI_[\u0649 \u0627]K*",
38
+ "26": "I_[\u0643]K*",
39
+ "27": "I_[ \u0644]K*",
40
+ "28": "I_[\u0644]K*",
41
+ "29": "I_[\u0645]K*",
42
+ "30": "I_[\u0646]K*",
43
+ "31": "I_[\u0647]K*",
44
+ "32": "I_[\u0648]K*",
45
+ "33": "I_[\u064a]K*",
46
+ "34": "I_[\u064a]KDK*",
47
+ "35": "K*",
48
+ "36": "K*A_[\u0623\u0646]",
49
+ "37": "K*A_[\u0639\u0644\u0649]",
50
+ "38": "K*A_[\u0641\u064a]",
51
+ "39": "K*A_[\u0645\u0646]",
52
+ "40": "K*D",
53
+ "41": "K*DD",
54
+ "42": "KDDK*",
55
+ "43": "K*DK",
56
+ "44": "KDK*",
57
+ "45": "K*DKI_[\u0630\u0627]",
58
+ "46": "K*DKK",
59
+ "47": "K*DKKK",
60
+ "48": "K*DKKKKKK",
61
+ "49": "KI_[ ]K*",
62
+ "50": "K*I_[\u0621]",
63
+ "51": "KI_[\u0623]K*",
64
+ "52": "KI_[\u0625]K*",
65
+ "53": "K*I_[\u0626]K",
66
+ "54": "K*I_[\u0626]KK",
67
+ "55": "K*I_[\u0626\u0627]",
68
+ "56": "K*I_[\u0627]",
69
+ "57": "K*I_[ \u0627]K",
70
+ "58": "K*I_[\u0627]K",
71
+ "59": "KI_[\u0627 ]K*",
72
+ "60": "K*I_[\u0627]KD",
73
+ "61": "K*I_[\u0627]KK",
74
+ "62": "K*I_[\u0627]KKK",
75
+ "63": "K*I_[\u0627\u062a]",
76
+ "64": "KI_[\u0627\u0644]K*",
77
+ "65": "K*I_[\u0627\u0646]",
78
+ "66": "K*I_[\u0629]",
79
+ "67": "K*I_[\u062a]",
80
+ "68": "K*I_[\u062a]K",
81
+ "69": "KI_[\u062a]K*",
82
+ "70": "K*I_[\u062a]KK",
83
+ "71": "K*I_[\u062c]KD",
84
+ "72": "K*I_[\u062f]",
85
+ "73": "K*I_[\u062f]K",
86
+ "74": "K*I_[\u0630]K",
87
+ "75": "K*I_[\u0630]KD",
88
+ "76": "KI_[\u0630]KI_[ \u0627]K*",
89
+ "77": "K*I_[\u0631]",
90
+ "78": "K*I_[\u0631]K",
91
+ "79": "K*I_[\u0643]K",
92
+ "80": "K*I_[\u0644]",
93
+ "81": "K*I_[\u0644]K",
94
+ "82": "KI_[\u0644]K*",
95
+ "83": "KI_[\u0644\u0623]K*",
96
+ "84": "K*I_[\u0644\u0627]",
97
+ "85": "K*I_[\u0644\u0649]",
98
+ "86": "KI_[\u0644\u0649 ]K*",
99
+ "87": "K*I_[\u0645]",
100
+ "88": "K*I_[\u0645]K",
101
+ "89": "K*I_[\u0645\u0627]",
102
+ "90": "K*I_[\u0646]",
103
+ "91": "K*I_[\u0646]K",
104
+ "92": "K*I_[\u0646]KK",
105
+ "93": "K*I_[\u0647]",
106
+ "94": "K*I_[\u0647]K",
107
+ "95": "K*I_[\u0647\u0627]",
108
+ "96": "K*I_[\u0648]K",
109
+ "97": "K*I_[\u0648]KK",
110
+ "98": "K*I_[\u0648\u0627]",
111
+ "99": "K*I_[\u0648\u0646]",
112
+ "100": "K*I_[\u0649]",
113
+ "101": "K*I_[\u064a]",
114
+ "102": "K*I_[\u064a]K",
115
+ "103": "KI_[\u064a ]K*",
116
+ "104": "KI_[\u064a]K*",
117
+ "105": "K*I_[\u064a]KI_[\u0627]",
118
+ "106": "K*I_[\u064a]KK",
119
+ "107": "K*I_[\u064a\u0627]K",
120
+ "108": "K*I_[\u064a\u0646]",
121
+ "109": "KKI_[ ]K*",
122
+ "110": "KKI_[ \u0623]K*",
123
+ "111": "KKI_[\u0623]K*",
124
+ "112": "KKI_[\u0625]K*",
125
+ "113": "KKI_[\u0644]K*",
126
+ "114": "KKI_[\u0645]K*",
127
+ "115": "KKKDK*",
128
+ "116": "KKKI_[ ]K*",
129
+ "117": "KKKKI_[ ]K*",
130
+ "118": "KKKR_[\u0623]K*",
131
+ "119": "KKKR_[\u0625]K*",
132
+ "120": "KKR_[\u0622]K*",
133
+ "121": "KKR_[\u0623]K*",
134
+ "122": "KKR_[\u0625]K*",
135
+ "123": "KKR_[\u0625]K*R_[\u064a]",
136
+ "124": "KKR_[\u0627]K*",
137
+ "125": "KKR_[\u062a]K*",
138
+ "126": "K*R_[ ]",
139
+ "127": "K*R_[\u0621]",
140
+ "128": "K*R_[\u0621]K",
141
+ "129": "K*R_[\u0621]KK",
142
+ "130": "K*R_[\u0622]",
143
+ "131": "K*R_[\u0622]K",
144
+ "132": "K*R_[\u0622]KK",
145
+ "133": "K*R_[\u0623]",
146
+ "134": "K*R_[\u0623]D",
147
+ "135": "KR_[\u0623]K*",
148
+ "136": "K*R_[\u0623]KD",
149
+ "137": "K*R_[\u0623]KI_[\u0647]",
150
+ "138": "K*R_[\u0623]KI_[\u064a]",
151
+ "139": "K*R_[\u0623]KKKKR_[\u064a]",
152
+ "140": "K*R_[\u0624]",
153
+ "141": "K*R_[\u0624]K",
154
+ "142": "K*R_[\u0624]KK",
155
+ "143": "K*R_[\u0625]",
156
+ "144": "K*R_[\u0625]I_[\u064a]K",
157
+ "145": "KR_[\u0625]K*",
158
+ "146": "K*R_[\u0626]",
159
+ "147": "K*R_[\u0626]I_[\u0627]",
160
+ "148": "K*R_[\u0626]K",
161
+ "149": "K*R_[\u0627]",
162
+ "150": "K*R_[\u0627]K",
163
+ "151": "K*R_[\u0627]KK",
164
+ "152": "K*R_[\u0628]",
165
+ "153": "K*R_[\u0629]",
166
+ "154": "K*R_[\u0629]D",
167
+ "155": "K*R_[\u062a]",
168
+ "156": "K*R_[\u062a]K",
169
+ "157": "KR_[\u062a]K*",
170
+ "158": "K*R_[\u062b]",
171
+ "159": "K*R_[\u062b]K",
172
+ "160": "K*R_[\u062b]KK",
173
+ "161": "K*R_[\u062c]",
174
+ "162": "K*R_[\u062c]K",
175
+ "163": "K*R_[\u062d]",
176
+ "164": "K*R_[\u062d]K",
177
+ "165": "K*R_[\u062f]",
178
+ "166": "K*R_[\u0630]",
179
+ "167": "K*R_[\u0630]D",
180
+ "168": "K*R_[\u0630]I_[\u0627]",
181
+ "169": "K*R_[\u0630]K",
182
+ "170": "K*R_[\u0630]KI_[\u0646]",
183
+ "171": "K*R_[\u0630]KK",
184
+ "172": "K*R_[\u0630]R_[\u064a]",
185
+ "173": "K*R_[\u0631]",
186
+ "174": "K*R_[\u0632]",
187
+ "175": "K*R_[\u0633]",
188
+ "176": "K*R_[\u0636]",
189
+ "177": "K*R_[\u0636]K",
190
+ "178": "K*R_[\u0638]",
191
+ "179": "K*R_[\u0639]",
192
+ "180": "K*R_[\u0641]",
193
+ "181": "K*R_[\u0642]",
194
+ "182": "K*R_[\u0642]K",
195
+ "183": "K*R_[\u0643]",
196
+ "184": "K*R_[\u0643]K",
197
+ "185": "K*R_[\u0644]",
198
+ "186": "K*R_[\u0644]I_[\u0649]",
199
+ "187": "K*R_[\u0645]",
200
+ "188": "K*R_[\u0646]",
201
+ "189": "KR_[\u0646]K*",
202
+ "190": "K*R_[\u0647]",
203
+ "191": "K*R_[\u0648]",
204
+ "192": "K*R_[\u0648]K",
205
+ "193": "K*R_[\u0649]",
206
+ "194": "K*R_[\u064a]",
207
+ "195": "K*R_[\u064a]I_[\u0621]",
208
+ "196": "K*R_[\u064a]I_[\u0627]",
209
+ "197": "K*R_[\u064a]K",
210
+ "198": "K*R_[\u064a]KK",
211
+ "199": "K*R_[\u064a]KKK",
212
+ "200": "MD*",
213
+ "201": "MDDK*",
214
+ "202": "MDK*",
215
+ "203": "MI_[\u0627\u0644]K*",
216
+ "204": "MI_[\u062a]K*",
217
+ "205": "MK*",
218
+ "206": "MK*D",
219
+ "207": "MK*DD",
220
+ "208": "MK*DK",
221
+ "209": "MK*I_[\u0627]",
222
+ "210": "MKKR_[\u0622]K*",
223
+ "211": "MKKR_[\u0623]K*",
224
+ "212": "MKKR_[\u0625]K*",
225
+ "213": "MKKR_[\u0627]K*",
226
+ "214": "MK*R_[\u0623]",
227
+ "215": "MK*R_[\u0623]K",
228
+ "216": "MK*R_[\u0627]",
229
+ "217": "MK*R_[\u0629]",
230
+ "218": "MK*R_[\u0649]",
231
+ "219": "MK*R_[\u064a]",
232
+ "220": "MR_[\u0622]K*",
233
+ "221": "MR_[\u0623]K*",
234
+ "222": "MR_[\u0625]K*",
235
+ "223": "MR_[\u0627]",
236
+ "224": "MR_[\u0627]K*",
237
+ "225": "MR_[\u062a]I_[\u0634]K*",
238
+ "226": "R_[ ]",
239
+ "227": "R_[0]",
240
+ "228": "R_[ ]D*",
241
+ "229": "R_[ ]I_[\u0623]K*",
242
+ "230": "R_[ ]I_[\u0625]K*",
243
+ "231": "R_[\u0621]",
244
+ "232": "R_[\u0621]K*",
245
+ "233": "R_[\u0622]K*",
246
+ "234": "R_[\u0623]",
247
+ "235": "R_[\u0623]D*",
248
+ "236": "R_[\u0623]DK*",
249
+ "237": "R_[\u0623]K*",
250
+ "238": "R_[\u0623]K*D",
251
+ "239": "R_[\u0623]K*I_[ ]K",
252
+ "240": "R_[\u0623]K*I_[\u0627]",
253
+ "241": "R_[\u0623]K*I_[\u062a]",
254
+ "242": "R_[\u0623]K*I_[\u0647]",
255
+ "243": "R_[\u0623]K*R_[\u0623]",
256
+ "244": "R_[\u0623]K*R_[\u062b]",
257
+ "245": "R_[\u0623]K*R_[\u0649]",
258
+ "246": "R_[\u0623]K*R_[\u064a]",
259
+ "247": "R_[\u0623]R_[\u0646]",
260
+ "248": "R_[\u0623]R_[\u064a]",
261
+ "249": "R_[\u0624]",
262
+ "250": "R_[\u0624]K*",
263
+ "251": "R_[\u0625]DK*",
264
+ "252": "R_[\u0625]I_[\u064a]K*",
265
+ "253": "R_[\u0625]K*",
266
+ "254": "R_[\u0625]K*D",
267
+ "255": "R_[\u0625]KI_[ ]K*",
268
+ "256": "R_[\u0625]K*I_[\u0627]",
269
+ "257": "R_[\u0625]K*R_[\u0649]",
270
+ "258": "R_[\u0626]",
271
+ "259": "R_[\u0626]K*",
272
+ "260": "R_[\u0627]",
273
+ "261": "R_[\u0627]K*",
274
+ "262": "R_[\u0628]",
275
+ "263": "R_[\u0628]D*",
276
+ "264": "R_[\u0628]DK*",
277
+ "265": "R_[\u0628]K*",
278
+ "266": "R_[\u0629]",
279
+ "267": "R_[\u0629]D*",
280
+ "268": "R_[\u0629]I_[ ]K*",
281
+ "269": "R_[\u062a]",
282
+ "270": "R_[\u062a]K*",
283
+ "271": "R_[\u062b]",
284
+ "272": "R_[\u062b]K*",
285
+ "273": "R_[\u062c]K*",
286
+ "274": "R_[\u062d]",
287
+ "275": "R_[\u062d]K*",
288
+ "276": "R_[\u062f]",
289
+ "277": "R_[\u062f]K*",
290
+ "278": "R_[\u0630]",
291
+ "279": "R_[\u0630]K*",
292
+ "280": "R_[\u0631]",
293
+ "281": "R_[\u0632]K*",
294
+ "282": "R_[\u0633]",
295
+ "283": "R_[\u0633]K*",
296
+ "284": "R_[\u0636]",
297
+ "285": "R_[\u0636]K*",
298
+ "286": "R_[\u0638]",
299
+ "287": "R_[\u0638]K*",
300
+ "288": "R_[\u0639]K*",
301
+ "289": "R_[\u063a]K*",
302
+ "290": "R_[\u0641]",
303
+ "291": "R_[\u0641]K*",
304
+ "292": "R_[\u0642]K*",
305
+ "293": "R_[\u0643]K*",
306
+ "294": "R_[\u0644]",
307
+ "295": "R_[\u0644]D*",
308
+ "296": "R_[\u0644]K*",
309
+ "297": "R_[\u0644]R_[\u0643]I_[\u0646]",
310
+ "298": "R_[\u0644]R_[\u064a]I_[\u0633]",
311
+ "299": "R_[\u0645]",
312
+ "300": "R_[\u0645]D*",
313
+ "301": "R_[\u0645]K*",
314
+ "302": "R_[\u0646]",
315
+ "303": "R_[\u0646]K*",
316
+ "304": "R_[\u0646]K*D",
317
+ "305": "R_[\u0647]",
318
+ "306": "R_[\u0647]K*",
319
+ "307": "R_[\u0647]R_[\u0646]I_[\u0627\u0643]",
320
+ "308": "R_[\u0648]",
321
+ "309": "R_[\u0648]K*",
322
+ "310": "R_[\u0649]",
323
+ "311": "R_[\u0649]D*",
324
+ "312": "R_[\u064a]",
325
+ "313": "R_[\u064a]I_[\u0621]",
326
+ "314": "R_[\u064a]K*"
327
+ },
328
+ "initializer_range": 0.02,
329
+ "intermediate_size": 3072,
330
+ "label2id": {
331
+ "D*": 0,
332
+ "DDK*": 1,
333
+ "DK*": 2,
334
+ "DKD*": 3,
335
+ "DKI_[\u0644]K*": 4,
336
+ "I_[ ]K*": 5,
337
+ "I_[ ]K*I_[ ]K": 6,
338
+ "I_[ ]K*I_[ ]KI_[ ]K": 7,
339
+ "I_[ ]K*R_[\u0623]": 8,
340
+ "I_[ \u0627]K*": 16,
341
+ "I_[ \u0644]K*": 27,
342
+ "I_[\u0623]K*": 9,
343
+ "I_[\u0623]K*D": 10,
344
+ "I_[\u0623\u0628]KD*": 11,
345
+ "I_[\u0625]K*": 12,
346
+ "I_[\u0625\u0630]KD*": 13,
347
+ "I_[\u0625\u0644]KD*": 14,
348
+ "I_[\u0626]K*": 15,
349
+ "I_[\u0627]K*": 17,
350
+ "I_[\u0627]KDK*": 18,
351
+ "I_[\u0627\u0644]K*": 19,
352
+ "I_[\u0628]K*": 20,
353
+ "I_[\u062a]K*": 21,
354
+ "I_[\u062f]K*": 22,
355
+ "I_[\u0631]K*": 23,
356
+ "I_[\u0639]K*": 24,
357
+ "I_[\u0639]KI_[\u0649 \u0627]K*": 25,
358
+ "I_[\u0643]K*": 26,
359
+ "I_[\u0644]K*": 28,
360
+ "I_[\u0645]K*": 29,
361
+ "I_[\u0646]K*": 30,
362
+ "I_[\u0647]K*": 31,
363
+ "I_[\u0648]K*": 32,
364
+ "I_[\u064a]K*": 33,
365
+ "I_[\u064a]KDK*": 34,
366
+ "K*": 35,
367
+ "K*A_[\u0623\u0646]": 36,
368
+ "K*A_[\u0639\u0644\u0649]": 37,
369
+ "K*A_[\u0641\u064a]": 38,
370
+ "K*A_[\u0645\u0646]": 39,
371
+ "K*D": 40,
372
+ "K*DD": 41,
373
+ "K*DK": 43,
374
+ "K*DKI_[\u0630\u0627]": 45,
375
+ "K*DKK": 46,
376
+ "K*DKKK": 47,
377
+ "K*DKKKKKK": 48,
378
+ "K*I_[ \u0627]K": 57,
379
+ "K*I_[\u0621]": 50,
380
+ "K*I_[\u0626]K": 53,
381
+ "K*I_[\u0626]KK": 54,
382
+ "K*I_[\u0626\u0627]": 55,
383
+ "K*I_[\u0627]": 56,
384
+ "K*I_[\u0627]K": 58,
385
+ "K*I_[\u0627]KD": 60,
386
+ "K*I_[\u0627]KK": 61,
387
+ "K*I_[\u0627]KKK": 62,
388
+ "K*I_[\u0627\u062a]": 63,
389
+ "K*I_[\u0627\u0646]": 65,
390
+ "K*I_[\u0629]": 66,
391
+ "K*I_[\u062a]": 67,
392
+ "K*I_[\u062a]K": 68,
393
+ "K*I_[\u062a]KK": 70,
394
+ "K*I_[\u062c]KD": 71,
395
+ "K*I_[\u062f]": 72,
396
+ "K*I_[\u062f]K": 73,
397
+ "K*I_[\u0630]K": 74,
398
+ "K*I_[\u0630]KD": 75,
399
+ "K*I_[\u0631]": 77,
400
+ "K*I_[\u0631]K": 78,
401
+ "K*I_[\u0643]K": 79,
402
+ "K*I_[\u0644]": 80,
403
+ "K*I_[\u0644]K": 81,
404
+ "K*I_[\u0644\u0627]": 84,
405
+ "K*I_[\u0644\u0649]": 85,
406
+ "K*I_[\u0645]": 87,
407
+ "K*I_[\u0645]K": 88,
408
+ "K*I_[\u0645\u0627]": 89,
409
+ "K*I_[\u0646]": 90,
410
+ "K*I_[\u0646]K": 91,
411
+ "K*I_[\u0646]KK": 92,
412
+ "K*I_[\u0647]": 93,
413
+ "K*I_[\u0647]K": 94,
414
+ "K*I_[\u0647\u0627]": 95,
415
+ "K*I_[\u0648]K": 96,
416
+ "K*I_[\u0648]KK": 97,
417
+ "K*I_[\u0648\u0627]": 98,
418
+ "K*I_[\u0648\u0646]": 99,
419
+ "K*I_[\u0649]": 100,
420
+ "K*I_[\u064a]": 101,
421
+ "K*I_[\u064a]K": 102,
422
+ "K*I_[\u064a]KI_[\u0627]": 105,
423
+ "K*I_[\u064a]KK": 106,
424
+ "K*I_[\u064a\u0627]K": 107,
425
+ "K*I_[\u064a\u0646]": 108,
426
+ "K*R_[ ]": 126,
427
+ "K*R_[\u0621]": 127,
428
+ "K*R_[\u0621]K": 128,
429
+ "K*R_[\u0621]KK": 129,
430
+ "K*R_[\u0622]": 130,
431
+ "K*R_[\u0622]K": 131,
432
+ "K*R_[\u0622]KK": 132,
433
+ "K*R_[\u0623]": 133,
434
+ "K*R_[\u0623]D": 134,
435
+ "K*R_[\u0623]KD": 136,
436
+ "K*R_[\u0623]KI_[\u0647]": 137,
437
+ "K*R_[\u0623]KI_[\u064a]": 138,
438
+ "K*R_[\u0623]KKKKR_[\u064a]": 139,
439
+ "K*R_[\u0624]": 140,
440
+ "K*R_[\u0624]K": 141,
441
+ "K*R_[\u0624]KK": 142,
442
+ "K*R_[\u0625]": 143,
443
+ "K*R_[\u0625]I_[\u064a]K": 144,
444
+ "K*R_[\u0626]": 146,
445
+ "K*R_[\u0626]I_[\u0627]": 147,
446
+ "K*R_[\u0626]K": 148,
447
+ "K*R_[\u0627]": 149,
448
+ "K*R_[\u0627]K": 150,
449
+ "K*R_[\u0627]KK": 151,
450
+ "K*R_[\u0628]": 152,
451
+ "K*R_[\u0629]": 153,
452
+ "K*R_[\u0629]D": 154,
453
+ "K*R_[\u062a]": 155,
454
+ "K*R_[\u062a]K": 156,
455
+ "K*R_[\u062b]": 158,
456
+ "K*R_[\u062b]K": 159,
457
+ "K*R_[\u062b]KK": 160,
458
+ "K*R_[\u062c]": 161,
459
+ "K*R_[\u062c]K": 162,
460
+ "K*R_[\u062d]": 163,
461
+ "K*R_[\u062d]K": 164,
462
+ "K*R_[\u062f]": 165,
463
+ "K*R_[\u0630]": 166,
464
+ "K*R_[\u0630]D": 167,
465
+ "K*R_[\u0630]I_[\u0627]": 168,
466
+ "K*R_[\u0630]K": 169,
467
+ "K*R_[\u0630]KI_[\u0646]": 170,
468
+ "K*R_[\u0630]KK": 171,
469
+ "K*R_[\u0630]R_[\u064a]": 172,
470
+ "K*R_[\u0631]": 173,
471
+ "K*R_[\u0632]": 174,
472
+ "K*R_[\u0633]": 175,
473
+ "K*R_[\u0636]": 176,
474
+ "K*R_[\u0636]K": 177,
475
+ "K*R_[\u0638]": 178,
476
+ "K*R_[\u0639]": 179,
477
+ "K*R_[\u0641]": 180,
478
+ "K*R_[\u0642]": 181,
479
+ "K*R_[\u0642]K": 182,
480
+ "K*R_[\u0643]": 183,
481
+ "K*R_[\u0643]K": 184,
482
+ "K*R_[\u0644]": 185,
483
+ "K*R_[\u0644]I_[\u0649]": 186,
484
+ "K*R_[\u0645]": 187,
485
+ "K*R_[\u0646]": 188,
486
+ "K*R_[\u0647]": 190,
487
+ "K*R_[\u0648]": 191,
488
+ "K*R_[\u0648]K": 192,
489
+ "K*R_[\u0649]": 193,
490
+ "K*R_[\u064a]": 194,
491
+ "K*R_[\u064a]I_[\u0621]": 195,
492
+ "K*R_[\u064a]I_[\u0627]": 196,
493
+ "K*R_[\u064a]K": 197,
494
+ "K*R_[\u064a]KK": 198,
495
+ "K*R_[\u064a]KKK": 199,
496
+ "KDDK*": 42,
497
+ "KDK*": 44,
498
+ "KI_[ ]K*": 49,
499
+ "KI_[\u0623]K*": 51,
500
+ "KI_[\u0625]K*": 52,
501
+ "KI_[\u0627 ]K*": 59,
502
+ "KI_[\u0627\u0644]K*": 64,
503
+ "KI_[\u062a]K*": 69,
504
+ "KI_[\u0630]KI_[ \u0627]K*": 76,
505
+ "KI_[\u0644]K*": 82,
506
+ "KI_[\u0644\u0623]K*": 83,
507
+ "KI_[\u0644\u0649 ]K*": 86,
508
+ "KI_[\u064a ]K*": 103,
509
+ "KI_[\u064a]K*": 104,
510
+ "KKI_[ ]K*": 109,
511
+ "KKI_[ \u0623]K*": 110,
512
+ "KKI_[\u0623]K*": 111,
513
+ "KKI_[\u0625]K*": 112,
514
+ "KKI_[\u0644]K*": 113,
515
+ "KKI_[\u0645]K*": 114,
516
+ "KKKDK*": 115,
517
+ "KKKI_[ ]K*": 116,
518
+ "KKKKI_[ ]K*": 117,
519
+ "KKKR_[\u0623]K*": 118,
520
+ "KKKR_[\u0625]K*": 119,
521
+ "KKR_[\u0622]K*": 120,
522
+ "KKR_[\u0623]K*": 121,
523
+ "KKR_[\u0625]K*": 122,
524
+ "KKR_[\u0625]K*R_[\u064a]": 123,
525
+ "KKR_[\u0627]K*": 124,
526
+ "KKR_[\u062a]K*": 125,
527
+ "KR_[\u0623]K*": 135,
528
+ "KR_[\u0625]K*": 145,
529
+ "KR_[\u062a]K*": 157,
530
+ "KR_[\u0646]K*": 189,
531
+ "MD*": 200,
532
+ "MDDK*": 201,
533
+ "MDK*": 202,
534
+ "MI_[\u0627\u0644]K*": 203,
535
+ "MI_[\u062a]K*": 204,
536
+ "MK*": 205,
537
+ "MK*D": 206,
538
+ "MK*DD": 207,
539
+ "MK*DK": 208,
540
+ "MK*I_[\u0627]": 209,
541
+ "MK*R_[\u0623]": 214,
542
+ "MK*R_[\u0623]K": 215,
543
+ "MK*R_[\u0627]": 216,
544
+ "MK*R_[\u0629]": 217,
545
+ "MK*R_[\u0649]": 218,
546
+ "MK*R_[\u064a]": 219,
547
+ "MKKR_[\u0622]K*": 210,
548
+ "MKKR_[\u0623]K*": 211,
549
+ "MKKR_[\u0625]K*": 212,
550
+ "MKKR_[\u0627]K*": 213,
551
+ "MR_[\u0622]K*": 220,
552
+ "MR_[\u0623]K*": 221,
553
+ "MR_[\u0625]K*": 222,
554
+ "MR_[\u0627]": 223,
555
+ "MR_[\u0627]K*": 224,
556
+ "MR_[\u062a]I_[\u0634]K*": 225,
557
+ "R_[ ]": 226,
558
+ "R_[ ]D*": 228,
559
+ "R_[ ]I_[\u0623]K*": 229,
560
+ "R_[ ]I_[\u0625]K*": 230,
561
+ "R_[0]": 227,
562
+ "R_[\u0621]": 231,
563
+ "R_[\u0621]K*": 232,
564
+ "R_[\u0622]K*": 233,
565
+ "R_[\u0623]": 234,
566
+ "R_[\u0623]D*": 235,
567
+ "R_[\u0623]DK*": 236,
568
+ "R_[\u0623]K*": 237,
569
+ "R_[\u0623]K*D": 238,
570
+ "R_[\u0623]K*I_[ ]K": 239,
571
+ "R_[\u0623]K*I_[\u0627]": 240,
572
+ "R_[\u0623]K*I_[\u062a]": 241,
573
+ "R_[\u0623]K*I_[\u0647]": 242,
574
+ "R_[\u0623]K*R_[\u0623]": 243,
575
+ "R_[\u0623]K*R_[\u062b]": 244,
576
+ "R_[\u0623]K*R_[\u0649]": 245,
577
+ "R_[\u0623]K*R_[\u064a]": 246,
578
+ "R_[\u0623]R_[\u0646]": 247,
579
+ "R_[\u0623]R_[\u064a]": 248,
580
+ "R_[\u0624]": 249,
581
+ "R_[\u0624]K*": 250,
582
+ "R_[\u0625]DK*": 251,
583
+ "R_[\u0625]I_[\u064a]K*": 252,
584
+ "R_[\u0625]K*": 253,
585
+ "R_[\u0625]K*D": 254,
586
+ "R_[\u0625]K*I_[\u0627]": 256,
587
+ "R_[\u0625]K*R_[\u0649]": 257,
588
+ "R_[\u0625]KI_[ ]K*": 255,
589
+ "R_[\u0626]": 258,
590
+ "R_[\u0626]K*": 259,
591
+ "R_[\u0627]": 260,
592
+ "R_[\u0627]K*": 261,
593
+ "R_[\u0628]": 262,
594
+ "R_[\u0628]D*": 263,
595
+ "R_[\u0628]DK*": 264,
596
+ "R_[\u0628]K*": 265,
597
+ "R_[\u0629]": 266,
598
+ "R_[\u0629]D*": 267,
599
+ "R_[\u0629]I_[ ]K*": 268,
600
+ "R_[\u062a]": 269,
601
+ "R_[\u062a]K*": 270,
602
+ "R_[\u062b]": 271,
603
+ "R_[\u062b]K*": 272,
604
+ "R_[\u062c]K*": 273,
605
+ "R_[\u062d]": 274,
606
+ "R_[\u062d]K*": 275,
607
+ "R_[\u062f]": 276,
608
+ "R_[\u062f]K*": 277,
609
+ "R_[\u0630]": 278,
610
+ "R_[\u0630]K*": 279,
611
+ "R_[\u0631]": 280,
612
+ "R_[\u0632]K*": 281,
613
+ "R_[\u0633]": 282,
614
+ "R_[\u0633]K*": 283,
615
+ "R_[\u0636]": 284,
616
+ "R_[\u0636]K*": 285,
617
+ "R_[\u0638]": 286,
618
+ "R_[\u0638]K*": 287,
619
+ "R_[\u0639]K*": 288,
620
+ "R_[\u063a]K*": 289,
621
+ "R_[\u0641]": 290,
622
+ "R_[\u0641]K*": 291,
623
+ "R_[\u0642]K*": 292,
624
+ "R_[\u0643]K*": 293,
625
+ "R_[\u0644]": 294,
626
+ "R_[\u0644]D*": 295,
627
+ "R_[\u0644]K*": 296,
628
+ "R_[\u0644]R_[\u0643]I_[\u0646]": 297,
629
+ "R_[\u0644]R_[\u064a]I_[\u0633]": 298,
630
+ "R_[\u0645]": 299,
631
+ "R_[\u0645]D*": 300,
632
+ "R_[\u0645]K*": 301,
633
+ "R_[\u0646]": 302,
634
+ "R_[\u0646]K*": 303,
635
+ "R_[\u0646]K*D": 304,
636
+ "R_[\u0647]": 305,
637
+ "R_[\u0647]K*": 306,
638
+ "R_[\u0647]R_[\u0646]I_[\u0627\u0643]": 307,
639
+ "R_[\u0648]": 308,
640
+ "R_[\u0648]K*": 309,
641
+ "R_[\u0649]": 310,
642
+ "R_[\u0649]D*": 311,
643
+ "R_[\u064a]": 312,
644
+ "R_[\u064a]I_[\u0621]": 313,
645
+ "R_[\u064a]K*": 314
646
+ },
647
+ "layer_norm_eps": 1e-12,
648
+ "max_position_embeddings": 512,
649
+ "model_type": "bert",
650
+ "num_attention_heads": 12,
651
+ "num_hidden_layers": 12,
652
+ "pad_token_id": 0,
653
+ "position_embedding_type": "absolute",
654
+ "torch_dtype": "float32",
655
+ "transformers_version": "4.30.0",
656
+ "type_vocab_size": 2,
657
+ "use_cache": true,
658
+ "vocab_size": 64000
659
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:584ccc089d143b1d7c72ea5b296652050359d163e57e4374b920fbac7925e8d6
3
+ size 539450353
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_basic_tokenize": true,
5
+ "do_lower_case": false,
6
+ "mask_token": "[MASK]",
7
+ "max_len": 512,
8
+ "model_max_length": 512,
9
+ "never_split": [
10
+ "[ุจุฑูŠุฏ]",
11
+ "[ู…ุณุชุฎุฏู…]",
12
+ "[ุฑุงุจุท]"
13
+ ],
14
+ "pad_token": "[PAD]",
15
+ "sep_token": "[SEP]",
16
+ "strip_accents": null,
17
+ "tokenize_chinese_chars": true,
18
+ "tokenizer_class": "BertTokenizer",
19
+ "unk_token": "[UNK]"
20
+ }
trainer_state.json ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 10.703363914373089,
5
+ "global_step": 7000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.76,
12
+ "learning_rate": 4.745158002038736e-05,
13
+ "loss": 0.3992,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 1.53,
18
+ "learning_rate": 4.490316004077472e-05,
19
+ "loss": 0.196,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 2.29,
24
+ "learning_rate": 4.235474006116208e-05,
25
+ "loss": 0.1557,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 3.06,
30
+ "learning_rate": 3.980632008154944e-05,
31
+ "loss": 0.1302,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 3.82,
36
+ "learning_rate": 3.72579001019368e-05,
37
+ "loss": 0.1053,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 4.59,
42
+ "learning_rate": 3.4709480122324164e-05,
43
+ "loss": 0.0898,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 5.35,
48
+ "learning_rate": 3.2161060142711516e-05,
49
+ "loss": 0.0762,
50
+ "step": 3500
51
+ },
52
+ {
53
+ "epoch": 6.12,
54
+ "learning_rate": 2.9612640163098882e-05,
55
+ "loss": 0.0644,
56
+ "step": 4000
57
+ },
58
+ {
59
+ "epoch": 6.88,
60
+ "learning_rate": 2.7064220183486238e-05,
61
+ "loss": 0.0536,
62
+ "step": 4500
63
+ },
64
+ {
65
+ "epoch": 7.65,
66
+ "learning_rate": 2.45158002038736e-05,
67
+ "loss": 0.0449,
68
+ "step": 5000
69
+ },
70
+ {
71
+ "epoch": 8.41,
72
+ "learning_rate": 2.196738022426096e-05,
73
+ "loss": 0.0391,
74
+ "step": 5500
75
+ },
76
+ {
77
+ "epoch": 9.17,
78
+ "learning_rate": 1.9418960244648318e-05,
79
+ "loss": 0.0331,
80
+ "step": 6000
81
+ },
82
+ {
83
+ "epoch": 9.94,
84
+ "learning_rate": 1.6870540265035677e-05,
85
+ "loss": 0.0289,
86
+ "step": 6500
87
+ },
88
+ {
89
+ "epoch": 10.7,
90
+ "learning_rate": 1.4322120285423038e-05,
91
+ "loss": 0.0248,
92
+ "step": 7000
93
+ }
94
+ ],
95
+ "max_steps": 9810,
96
+ "num_train_epochs": 15,
97
+ "total_flos": 2.6176216075199188e+16,
98
+ "trial_name": null,
99
+ "trial_params": null
100
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0034271e59af5431f3710990a7edf92ec80e6bc4da6b7128555fd6809e0ae32
3
+ size 4143
vocab.txt ADDED
The diff for this file is too large to render. See raw diff