Upload MemoryLLM
Browse files- README.md +2 -9
- config.json +10 -3
- model-00001-of-00008.safetensors +1 -1
- model-00002-of-00008.safetensors +1 -1
- model-00003-of-00008.safetensors +1 -1
- model-00004-of-00008.safetensors +1 -1
- model-00005-of-00008.safetensors +1 -1
- model-00006-of-00008.safetensors +1 -1
- model-00007-of-00008.safetensors +1 -1
- model-00008-of-00008.safetensors +1 -1
README.md
CHANGED
@@ -40,15 +40,8 @@ messages = [{
|
|
40 |
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
|
41 |
inputs = inputs[:, 1:] # remove bos token
|
42 |
|
43 |
-
|
44 |
-
|
45 |
-
tokenizer.convert_tokens_to_ids("<|eot_id|>")
|
46 |
-
]
|
47 |
-
|
48 |
-
outputs = model.generate(input_ids=inputs.cuda()[:, 1:],
|
49 |
-
max_new_tokens=20,
|
50 |
-
eos_token_id=terminators)
|
51 |
-
|
52 |
response = tokenizer.decode(outputs[0])
|
53 |
|
54 |
outputs = model.generate(inputs=input_ids.cuda(), attention_mask=attention_mask.cuda(), max_new_tokens=10)
|
|
|
40 |
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
|
41 |
inputs = inputs[:, 1:] # remove bos token
|
42 |
|
43 |
+
outputs = model.generate(input_ids=inputs.cuda(),
|
44 |
+
max_new_tokens=20)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
response = tokenizer.decode(outputs[0])
|
46 |
|
47 |
outputs = model.generate(inputs=input_ids.cuda(), attention_mask=attention_mask.cuda(), max_new_tokens=10)
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "memoryllm-8b-chat",
|
3 |
"add_bos_embedding": true,
|
4 |
"add_decoder_lora": true,
|
5 |
"architectures": [
|
@@ -25,7 +25,8 @@
|
|
25 |
"k_proj",
|
26 |
"up_proj",
|
27 |
"down_proj",
|
28 |
-
"gate_proj"
|
|
|
29 |
]
|
30 |
},
|
31 |
"max_length": 512,
|
@@ -40,7 +41,7 @@
|
|
40 |
"num_key_value_heads": 8,
|
41 |
"num_memory_tokens": 12800,
|
42 |
"num_tokens": 256,
|
43 |
-
"pad_token_id":
|
44 |
"pretraining_tp": 1,
|
45 |
"rms_norm_eps": 1e-05,
|
46 |
"rope_scaling": {
|
@@ -49,9 +50,15 @@
|
|
49 |
},
|
50 |
"rope_theta": 500000.0,
|
51 |
"shrink_to_one_embedding": true,
|
|
|
|
|
|
|
|
|
|
|
52 |
"tie_word_embeddings": false,
|
53 |
"torch_dtype": "float32",
|
54 |
"transformers_version": "4.43.0.dev0",
|
|
|
55 |
"use_cache": true,
|
56 |
"vocab_size": 128256
|
57 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "locals/memoryllm-8b-chat",
|
3 |
"add_bos_embedding": true,
|
4 |
"add_decoder_lora": true,
|
5 |
"architectures": [
|
|
|
25 |
"k_proj",
|
26 |
"up_proj",
|
27 |
"down_proj",
|
28 |
+
"gate_proj",
|
29 |
+
"lm_head"
|
30 |
]
|
31 |
},
|
32 |
"max_length": 512,
|
|
|
41 |
"num_key_value_heads": 8,
|
42 |
"num_memory_tokens": 12800,
|
43 |
"num_tokens": 256,
|
44 |
+
"pad_token_id": 128009,
|
45 |
"pretraining_tp": 1,
|
46 |
"rms_norm_eps": 1e-05,
|
47 |
"rope_scaling": {
|
|
|
50 |
},
|
51 |
"rope_theta": 500000.0,
|
52 |
"shrink_to_one_embedding": true,
|
53 |
+
"special_token_ids": [
|
54 |
+
128006,
|
55 |
+
128007,
|
56 |
+
128009
|
57 |
+
],
|
58 |
"tie_word_embeddings": false,
|
59 |
"torch_dtype": "float32",
|
60 |
"transformers_version": "4.43.0.dev0",
|
61 |
+
"tune_special_tokens": true,
|
62 |
"use_cache": true,
|
63 |
"vocab_size": 128256
|
64 |
}
|
model-00001-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6710886520
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7af65cce95946eeee8693d00b8a1d6695de24f3801256061f769082ec3bebe2f
|
3 |
size 6710886520
|
model-00002-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4917689241
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3ec2e5cccec05a70c3a968b4a3c8a0f05f49aa028aad8218115e2b4100e90b6
|
3 |
size 4917689241
|
model-00003-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4883929088
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:116b96cb823882cfe86460aa84c99ca793deafafe8c40590f37e7e0a8412be45
|
3 |
size 4883929088
|
model-00004-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4816854808
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e17b030cf85a2824f567c948f3e269a836092307c9dde1857063fdc3891b872
|
3 |
size 4816854808
|
model-00005-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4883929264
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e32e3065fa8d4716745ea8a5427ba20467f591fd06df27d120a93a1e41223805
|
3 |
size 4883929264
|
model-00006-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4816854856
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8efe66af244af0fadf5a3ab3dd080f2ff7311a65653179251c9f5ee204b61ba4
|
3 |
size 4816854856
|
model-00007-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4883929264
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e44fc390f8a968a6534b3369b5d4387a60cad4204c5a6a5e9ac0b46c7c19a308
|
3 |
size 4883929264
|
model-00008-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3220526344
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcf4b278965fd94af150585d96f3051c1c1dada3b54246757b78dbf6ad8d58b3
|
3 |
size 3220526344
|