MohametSena commited on
Commit
2bc0069
·
1 Parent(s): 70a415c

End of training

Browse files
config.json CHANGED
@@ -15,5 +15,5 @@
15
  "LABEL_0": 0
16
  },
17
  "torch_dtype": "float32",
18
- "transformers_version": "4.27.0.dev0"
19
  }
 
15
  "LABEL_0": 0
16
  },
17
  "torch_dtype": "float32",
18
+ "transformers_version": "4.29.1"
19
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d5f0487a4a40f6eac52e612c617152ed4ada7c154f75e8a2bd90f33ed055eb8
3
  size 713857157
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4222a3a26b6a317f1a1a69081c520b8039bec8ed6c3adebd272bdf6ebb6fefda
3
  size 713857157
tokenizer_config.json CHANGED
@@ -1,11 +1,11 @@
1
  {
 
2
  "cls_token": "[CLS]",
3
  "do_lower_case": false,
4
  "mask_token": "[MASK]",
5
  "model_max_length": 512,
6
  "pad_token": "[PAD]",
7
  "sep_token": "[SEP]",
8
- "special_tokens_map_file": null,
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
  "tokenizer_class": "BertTokenizer",
 
1
  {
2
+ "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
  "do_lower_case": false,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 512,
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
 
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
  "tokenizer_class": "BertTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e77ee31f450e5e9a42e2c782b848398201db9063d59e0262aa6ef1fe296ee9e
3
- size 3515
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f36adeb27b6ca307cbc5decae1f42cb2d22273cd1fa6ef0f4cdea5ea987494ee
3
+ size 3899