Training in progress, epoch 8
Browse files- model.safetensors +1 -1
- tokenizer.json +7 -12
- tokenizer_config.json +2 -2
- training_args.bin +2 -2
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 197603580
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:68d72d3c70e567e4a10d0457372572ffdfa25baf26e42c939e6554a8981c242a
|
3 |
size 197603580
|
tokenizer.json
CHANGED
@@ -1,11 +1,6 @@
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
-
"truncation":
|
4 |
-
"direction": "Right",
|
5 |
-
"max_length": 128,
|
6 |
-
"strategy": "LongestFirst",
|
7 |
-
"stride": 0
|
8 |
-
},
|
9 |
"padding": null,
|
10 |
"added_tokens": [
|
11 |
{
|
@@ -303,7 +298,7 @@
|
|
303 |
},
|
304 |
{
|
305 |
"SpecialToken": {
|
306 |
-
"id": "
|
307 |
"type_id": 0
|
308 |
}
|
309 |
}
|
@@ -329,7 +324,7 @@
|
|
329 |
},
|
330 |
{
|
331 |
"SpecialToken": {
|
332 |
-
"id": "
|
333 |
"type_id": 0
|
334 |
}
|
335 |
}
|
@@ -344,13 +339,13 @@
|
|
344 |
"</s>"
|
345 |
]
|
346 |
},
|
347 |
-
"
|
348 |
-
"id": "
|
349 |
"ids": [
|
350 |
-
|
351 |
],
|
352 |
"tokens": [
|
353 |
-
"
|
354 |
]
|
355 |
}
|
356 |
}
|
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
+
"truncation": null,
|
|
|
|
|
|
|
|
|
|
|
4 |
"padding": null,
|
5 |
"added_tokens": [
|
6 |
{
|
|
|
298 |
},
|
299 |
{
|
300 |
"SpecialToken": {
|
301 |
+
"id": "hi_IN",
|
302 |
"type_id": 0
|
303 |
}
|
304 |
}
|
|
|
324 |
},
|
325 |
{
|
326 |
"SpecialToken": {
|
327 |
+
"id": "hi_IN",
|
328 |
"type_id": 0
|
329 |
}
|
330 |
}
|
|
|
339 |
"</s>"
|
340 |
]
|
341 |
},
|
342 |
+
"hi_IN": {
|
343 |
+
"id": "hi_IN",
|
344 |
"ids": [
|
345 |
+
65545
|
346 |
],
|
347 |
"tokens": [
|
348 |
+
"hi_IN"
|
349 |
]
|
350 |
}
|
351 |
}
|
tokenizer_config.json
CHANGED
@@ -279,8 +279,8 @@
|
|
279 |
"model_max_length": 1024,
|
280 |
"pad_token": "<pad>",
|
281 |
"sep_token": "</s>",
|
282 |
-
"src_lang": "
|
283 |
-
"tgt_lang": "
|
284 |
"tokenizer_class": "MBartTokenizer",
|
285 |
"trim_offsets": true,
|
286 |
"unk_token": "<unk>"
|
|
|
279 |
"model_max_length": 1024,
|
280 |
"pad_token": "<pad>",
|
281 |
"sep_token": "</s>",
|
282 |
+
"src_lang": "hi_IN",
|
283 |
+
"tgt_lang": "hi_IN",
|
284 |
"tokenizer_class": "MBartTokenizer",
|
285 |
"trim_offsets": true,
|
286 |
"unk_token": "<unk>"
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b9ab7eba83a79e87db243724a12249417762143b24d1e6428dc9f11c29fa98ab
|
3 |
+
size 5368
|