YuWangX commited on
Commit
e09761e
·
verified ·
1 Parent(s): 4da8030

Upload MemoryLLM

Browse files
README.md CHANGED
@@ -40,15 +40,8 @@ messages = [{
40
  inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
41
  inputs = inputs[:, 1:] # remove bos token
42
 
43
- terminators = [
44
- tokenizer.eos_token_id,
45
- tokenizer.convert_tokens_to_ids("<|eot_id|>")
46
- ]
47
-
48
- outputs = model.generate(input_ids=inputs.cuda()[:, 1:],
49
- max_new_tokens=20,
50
- eos_token_id=terminators)
51
-
52
  response = tokenizer.decode(outputs[0])
53
 
54
  outputs = model.generate(inputs=input_ids.cuda(), attention_mask=attention_mask.cuda(), max_new_tokens=10)
 
40
  inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
41
  inputs = inputs[:, 1:] # remove bos token
42
 
43
+ outputs = model.generate(input_ids=inputs.cuda(),
44
+ max_new_tokens=20)
 
 
 
 
 
 
 
45
  response = tokenizer.decode(outputs[0])
46
 
47
  outputs = model.generate(inputs=input_ids.cuda(), attention_mask=attention_mask.cuda(), max_new_tokens=10)
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "memoryllm-8b-chat",
3
  "add_bos_embedding": true,
4
  "add_decoder_lora": true,
5
  "architectures": [
@@ -25,7 +25,8 @@
25
  "k_proj",
26
  "up_proj",
27
  "down_proj",
28
- "gate_proj"
 
29
  ]
30
  },
31
  "max_length": 512,
@@ -40,7 +41,7 @@
40
  "num_key_value_heads": 8,
41
  "num_memory_tokens": 12800,
42
  "num_tokens": 256,
43
- "pad_token_id": 128001,
44
  "pretraining_tp": 1,
45
  "rms_norm_eps": 1e-05,
46
  "rope_scaling": {
@@ -49,9 +50,15 @@
49
  },
50
  "rope_theta": 500000.0,
51
  "shrink_to_one_embedding": true,
 
 
 
 
 
52
  "tie_word_embeddings": false,
53
  "torch_dtype": "float32",
54
  "transformers_version": "4.43.0.dev0",
 
55
  "use_cache": true,
56
  "vocab_size": 128256
57
  }
 
1
  {
2
+ "_name_or_path": "locals/memoryllm-8b-chat",
3
  "add_bos_embedding": true,
4
  "add_decoder_lora": true,
5
  "architectures": [
 
25
  "k_proj",
26
  "up_proj",
27
  "down_proj",
28
+ "gate_proj",
29
+ "lm_head"
30
  ]
31
  },
32
  "max_length": 512,
 
41
  "num_key_value_heads": 8,
42
  "num_memory_tokens": 12800,
43
  "num_tokens": 256,
44
+ "pad_token_id": 128009,
45
  "pretraining_tp": 1,
46
  "rms_norm_eps": 1e-05,
47
  "rope_scaling": {
 
50
  },
51
  "rope_theta": 500000.0,
52
  "shrink_to_one_embedding": true,
53
+ "special_token_ids": [
54
+ 128006,
55
+ 128007,
56
+ 128009
57
+ ],
58
  "tie_word_embeddings": false,
59
  "torch_dtype": "float32",
60
  "transformers_version": "4.43.0.dev0",
61
+ "tune_special_tokens": true,
62
  "use_cache": true,
63
  "vocab_size": 128256
64
  }
model-00001-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2713410b4c31a0923361c8e21553a13a43c5af08aacca91305d8bfd555f86730
3
  size 6710886520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7af65cce95946eeee8693d00b8a1d6695de24f3801256061f769082ec3bebe2f
3
  size 6710886520
model-00002-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d5edfd6ebf9e3bcadf6143f8d7924f87409842b9210a00b3a92615f9ee1b7fa
3
  size 4917689241
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3ec2e5cccec05a70c3a968b4a3c8a0f05f49aa028aad8218115e2b4100e90b6
3
  size 4917689241
model-00003-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0dae25919a3894e7ecf58331c96d83ba1b6808337de473bc2e7bd1d98c863cb
3
  size 4883929088
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:116b96cb823882cfe86460aa84c99ca793deafafe8c40590f37e7e0a8412be45
3
  size 4883929088
model-00004-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13b18147bef5aa85bc9bcc820893862c9803d1da156b71019c19b3bd2e97df57
3
  size 4816854808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e17b030cf85a2824f567c948f3e269a836092307c9dde1857063fdc3891b872
3
  size 4816854808
model-00005-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:58733b1be1d8b93ec09c833ebd28e70a9e4ef3f2e742215b7e68cf5edfab294a
3
  size 4883929264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e32e3065fa8d4716745ea8a5427ba20467f591fd06df27d120a93a1e41223805
3
  size 4883929264
model-00006-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50b8f69fc77bb9b0dee767b56100d10bacea05882cb15e4aa92edf9949034a86
3
  size 4816854856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8efe66af244af0fadf5a3ab3dd080f2ff7311a65653179251c9f5ee204b61ba4
3
  size 4816854856
model-00007-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:410de571515c3fd957296da3d6a1e1579820f575fef227ce5517496f5d4e162c
3
  size 4883929264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e44fc390f8a968a6534b3369b5d4387a60cad4204c5a6a5e9ac0b46c7c19a308
3
  size 4883929264
model-00008-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a90c176b152a64e78d5b756b9562ac1a251009ce70bb766b3de43f5232aad9f
3
  size 3220526344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcf4b278965fd94af150585d96f3051c1c1dada3b54246757b78dbf6ad8d58b3
3
  size 3220526344