MohametSena commited on
Commit
da41d17
·
1 Parent(s): 48b3272

Training in progress, step 3300

Browse files
config.json CHANGED
@@ -15,5 +15,5 @@
15
  "LABEL_0": 0
16
  },
17
  "torch_dtype": "float32",
18
- "transformers_version": "4.29.1"
19
  }
 
15
  "LABEL_0": 0
16
  },
17
  "torch_dtype": "float32",
18
+ "transformers_version": "4.27.0.dev0"
19
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c91713406f6ec0a5c2c51fc0172d9fa84b6958a1fc7208a42fffa5dd223f1527
3
  size 713857157
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4edb3444ed8acb9f91f148d372db9666275aecff55ceaa67a8cca2ee7237c69b
3
  size 713857157
tokenizer_config.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
- "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
  "do_lower_case": false,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 512,
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
 
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
  "tokenizer_class": "BertTokenizer",
 
1
  {
 
2
  "cls_token": "[CLS]",
3
  "do_lower_case": false,
4
  "mask_token": "[MASK]",
5
  "model_max_length": 512,
6
  "pad_token": "[PAD]",
7
  "sep_token": "[SEP]",
8
+ "special_tokens_map_file": null,
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
  "tokenizer_class": "BertTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cf886a1c6dfa885cbd6e3d8e636466343b9c17cfa5821cbb3111e828b2d76bd
3
- size 3899
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:094d80a7fa2680efa9c51ac62f1ff84ab09dc11d0f1d99212296a127f411ca04
3
+ size 3515