at676 commited on
Commit
18caa0f
·
verified ·
1 Parent(s): ca8cedf

Upload folder using huggingface_hub (#2)

Browse files

- de3324ee8c51485a0c06a685dc60b1a97ff04811fc3f3a5ea48028889d4d6dd1 (b341d10a84275d4cc8c057dafe312bfd3a37c604)
- b984fb8327272ba41101640a28bcb1e1c22679fb99ee0f04365b280cfd53ff11 (b4a4f8a98f7458d0d3cb7565cc6e84de95f61549)
- 71cbe7f51cccae39f9fc8724e44c9912133822354767f54d42a4134b51e960d9 (bd815ebfe6af26140bdda185a9a558551692023b)
- 23886541a00488030342aa9ecb69773dd97116f57de55f730ff91c079279cdb5 (41f17cc0a6927d99d6383ad2597f1aa8d681e770)
- 618acca736657a93b554d3fa5958bd7866aab51183284daea6f8266dce323185 (90df972b2020ade46aa089addb05c4d4f018d277)
- 16d3f7b9fdd2480645e4a95a1784a2fb11721e3f36ad8a3151af5c0977185f28 (7583cda18cb7b16f6b66a8790f9d0ef8b8435453)

config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "meta-llama/Llama-2-70b-chat-hf",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 8192,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 28672,
14
+ "max_position_embeddings": 4096,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 64,
18
+ "num_hidden_layers": 80,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "quip_params": {
22
+ "K": 3,
23
+ "L": 16,
24
+ "V": 2,
25
+ "codebook": "bitshift",
26
+ "codebook_version": 0,
27
+ "decode_mode": "quantlut_sym",
28
+ "td_x": 16,
29
+ "td_y": 16,
30
+ "tlut_bits": 9
31
+ },
32
+ "rms_norm_eps": 1e-05,
33
+ "rope_scaling": null,
34
+ "rope_theta": 10000.0,
35
+ "tie_word_embeddings": false,
36
+ "torch_dtype": "float16",
37
+ "transformers_version": "4.44.2",
38
+ "use_cache": true,
39
+ "vocab_size": 32000
40
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "do_sample": true,
4
+ "eos_token_id": 2,
5
+ "max_length": 4096,
6
+ "pad_token_id": 0,
7
+ "temperature": 0.6,
8
+ "top_p": 0.9,
9
+ "transformers_version": "4.44.2"
10
+ }
model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73a87741c697f41bf4334c89af0669f002688884710d6c1316e368a8c3067e79
3
+ size 4935769496
model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bdf8541f4a4e28c20a340f43b3e8d5bfc4fe8e7413a3eb3ad1b130d10be1b36
3
+ size 4966068848
model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9af50df168f5cfc8c623f99e6b09b1576ed5b3d80640116b6da3b77322ab0552
3
+ size 4997518464
model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfdb74cfd1ed08c6c74bee985e8f3dcb4663a7148c5fd0f2662fe0ac5208d67f
3
+ size 4966068848
model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99af577269f13a0318b67fd406b3217eea9266ab1788756221fa2e163eae2245
3
+ size 4997518464
model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ca2b49ae7cd5b00d3f01c604a83d41e301365c5349b104d2d925404ec002a5a
3
+ size 1898132784
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff