almatkai commited on
Commit
d1692d2
·
1 Parent(s): 587be62
config.json CHANGED
@@ -1,8 +1,5 @@
1
  {
2
- "_name_or_path": "togethercomputer/Mistral-7B-Instruct-v0.1",
3
- "architectures": [
4
- "MistralForCausalLM"
5
- ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
@@ -15,12 +12,12 @@
15
  "num_attention_heads": 32,
16
  "num_hidden_layers": 32,
17
  "num_key_value_heads": 8,
18
- "rms_norm_eps": 1e-05,
19
  "rope_theta": 10000.0,
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
- "torch_dtype": "bfloat16",
23
- "transformers_version": "4.37.0.dev0",
24
- "use_cache": false,
25
  "vocab_size": 32000
26
  }
 
1
  {
2
+ "architectures": ["MistralForCausalLM"],
 
 
 
3
  "attention_dropout": 0.0,
4
  "bos_token_id": 1,
5
  "eos_token_id": 2,
 
12
  "num_attention_heads": 32,
13
  "num_hidden_layers": 32,
14
  "num_key_value_heads": 8,
15
+ "rms_norm_eps": 1e-5,
16
  "rope_theta": 10000.0,
17
  "sliding_window": 4096,
18
  "tie_word_embeddings": false,
19
+ "torch_dtype": "float16",
20
+ "transformers_version": "4.36.1",
21
+ "use_cache": true,
22
  "vocab_size": 32000
23
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.37.0.dev0"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.36.1"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d1cecb83a2dd6d9b471e741e1a9b90e9411efaf411ac33b1b07024e469d5b96d
3
- size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:749b14e942ec7cb8da86dce9a8f028b9392d6eddea206dfe726d9da9d0a403c7
3
+ size 4943162240
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d206dc2c5cdc68a98a61feb1936933a2bf6cb6e5a6095500bd278ba604408d9
3
- size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f07d8b4a4813bb39db98c3e9944d56d2fac2062fbd060ea0e693b52aa12dcb3
3
+ size 4999819232
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:244648d615134a5785da3a6ecc1a48e0e8727da53faa70c337f61a4a9c31d721
3
- size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc146dd49bb5d5201132f93cd310bf8260946b711516d4174a72162fe79fdd1e
3
+ size 4540516256
special_tokens_map.json CHANGED
@@ -13,7 +13,6 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "</s>",
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
16
  "unk_token": {
17
  "content": "<unk>",
18
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -32,8 +32,8 @@
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "legacy": true,
35
- "model_max_length": 8192,
36
- "pad_token": "</s>",
37
  "sp_model_kwargs": {},
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",
 
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "legacy": true,
35
+ "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": null,
37
  "sp_model_kwargs": {},
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",