Tarek07 commited on
Commit
332c697
·
verified ·
1 Parent(s): c7ef24b

Add files using upload-large-folder tool

Browse files
.ipynb_checkpoints/mergekit_config-checkpoint.yml ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - model: downloads/L3.1-70B-Hanami-x1
3
+ - model: downloads/70B-L3.3-Cirrus-x1
4
+ - model: downloads/Wayfarer-Large-70B-Llama-3.3
5
+ - model: downloads/Negative_LLAMA_70B
6
+ - model: downloads/Anubis-70B-v1
7
+ - model: downloads/EVA-LLaMA-3.33-70B-v0.1
8
+ merge_method: sce
9
+ base_model: downloads/Llama-3.1-Nemotron-lorablated-70B
10
+ parameters:
11
+ select_topk: 0.17
12
+ out_dtype: bfloat16
13
+ tokenizer:
14
+ source: downloads/Negative_LLAMA_70B
README.md CHANGED
@@ -1,34 +1,29 @@
1
  ---
2
- base_model:
3
- - EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
4
- - nbeerbower/Llama-3.1-Nemotron-lorablated-70B
5
- - Sao10K/L3.1-70B-Hanami-x1
6
- - SicariusSicariiStuff/Negative_LLAMA_70B
7
- - TheDrummer/Anubis-70B-v1
8
- - LatitudeGames/Wayfarer-Large-70B-Llama-3.3
9
  library_name: transformers
10
  tags:
11
  - mergekit
12
  - merge
13
 
14
  ---
15
- # merge
16
 
17
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
 
19
  ## Merge Details
20
  ### Merge Method
21
 
22
- This model was merged using the [Linear DELLA](https://arxiv.org/abs/2406.11617) merge method using [nbeerbower/Llama-3.1-Nemotron-lorablated-70B](https://huggingface.co/nbeerbower/Llama-3.1-Nemotron-lorablated-70B) as a base.
23
 
24
  ### Models Merged
25
 
26
  The following models were included in the merge:
27
- * [EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1](https://huggingface.co/EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1)
28
- * [Sao10K/L3.1-70B-Hanami-x1](https://huggingface.co/Sao10K/L3.1-70B-Hanami-x1)
29
- * [SicariusSicariiStuff/Negative_LLAMA_70B](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B)
30
- * [TheDrummer/Anubis-70B-v1](https://huggingface.co/TheDrummer/Anubis-70B-v1)
31
- * [LatitudeGames/Wayfarer-Large-70B-Llama-3.3](https://huggingface.co/LatitudeGames/Wayfarer-Large-70B-Llama-3.3)
 
32
 
33
  ### Configuration
34
 
@@ -36,33 +31,17 @@ The following YAML configuration was used to produce this model:
36
 
37
  ```yaml
38
  models:
39
- - model: Sao10K/L3.1-70B-Hanami-x1
40
- parameters:
41
- weight: 0.20
42
- density: 0.7
43
- - model: LatitudeGames/Wayfarer-Large-70B-Llama-3.3
44
- parameters:
45
- weight: 0.20
46
- density: 0.7
47
- - model: SicariusSicariiStuff/Negative_LLAMA_70B
48
- parameters:
49
- weight: 0.20
50
- density: 0.7
51
- - model: TheDrummer/Anubis-70B-v1
52
- parameters:
53
- weight: 0.20
54
- density: 0.7
55
- - model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
56
- parameters:
57
- weight: 0.20
58
- density: 0.7
59
- merge_method: della_linear
60
- base_model: nbeerbower/Llama-3.1-Nemotron-lorablated-70B
61
  parameters:
62
- epsilon: 0.2
63
- lambda: 1.1
64
  out_dtype: bfloat16
65
  tokenizer:
66
- source: union
67
-
68
  ```
 
1
  ---
2
+ base_model: []
 
 
 
 
 
 
3
  library_name: transformers
4
  tags:
5
  - mergekit
6
  - merge
7
 
8
  ---
9
+ # PrimogentorSCE3
10
 
11
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
12
 
13
  ## Merge Details
14
  ### Merge Method
15
 
16
+ This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using downloads/Llama-3.1-Nemotron-lorablated-70B as a base.
17
 
18
  ### Models Merged
19
 
20
  The following models were included in the merge:
21
+ * downloads/Wayfarer-Large-70B-Llama-3.3
22
+ * downloads/EVA-LLaMA-3.33-70B-v0.1
23
+ * downloads/Negative_LLAMA_70B
24
+ * downloads/L3.1-70B-Hanami-x1
25
+ * downloads/70B-L3.3-Cirrus-x1
26
+ * downloads/Anubis-70B-v1
27
 
28
  ### Configuration
29
 
 
31
 
32
  ```yaml
33
  models:
34
+ - model: downloads/L3.1-70B-Hanami-x1
35
+ - model: downloads/70B-L3.3-Cirrus-x1
36
+ - model: downloads/Wayfarer-Large-70B-Llama-3.3
37
+ - model: downloads/Negative_LLAMA_70B
38
+ - model: downloads/Anubis-70B-v1
39
+ - model: downloads/EVA-LLaMA-3.33-70B-v0.1
40
+ merge_method: sce
41
+ base_model: downloads/Llama-3.1-Nemotron-lorablated-70B
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  parameters:
43
+ select_topk: 0.17
 
44
  out_dtype: bfloat16
45
  tokenizer:
46
+ source: downloads/Negative_LLAMA_70B
 
47
  ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "nbeerbower/Llama-3.1-Nemotron-lorablated-70B",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -34,7 +34,7 @@
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "bfloat16",
37
- "transformers_version": "4.48.2",
38
  "use_cache": true,
39
  "vocab_size": 128256
40
  }
 
1
  {
2
+ "_name_or_path": "downloads/Llama-3.1-Nemotron-lorablated-70B",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "bfloat16",
37
+ "transformers_version": "4.49.0",
38
  "use_cache": true,
39
  "vocab_size": 128256
40
  }
mergekit_config.yml CHANGED
@@ -1,29 +1,14 @@
1
  models:
2
- - model: Sao10K/L3.1-70B-Hanami-x1
3
- parameters:
4
- weight: 0.20
5
- density: 0.7
6
- - model: LatitudeGames/Wayfarer-Large-70B-Llama-3.3
7
- parameters:
8
- weight: 0.20
9
- density: 0.7
10
- - model: SicariusSicariiStuff/Negative_LLAMA_70B
11
- parameters:
12
- weight: 0.20
13
- density: 0.7
14
- - model: TheDrummer/Anubis-70B-v1
15
- parameters:
16
- weight: 0.20
17
- density: 0.7
18
- - model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
19
- parameters:
20
- weight: 0.20
21
- density: 0.7
22
- merge_method: della_linear
23
- base_model: nbeerbower/Llama-3.1-Nemotron-lorablated-70B
24
  parameters:
25
- epsilon: 0.2
26
- lambda: 1.1
27
  out_dtype: bfloat16
28
  tokenizer:
29
- source: union
 
1
  models:
2
+ - model: downloads/L3.1-70B-Hanami-x1
3
+ - model: downloads/70B-L3.3-Cirrus-x1
4
+ - model: downloads/Wayfarer-Large-70B-Llama-3.3
5
+ - model: downloads/Negative_LLAMA_70B
6
+ - model: downloads/Anubis-70B-v1
7
+ - model: downloads/EVA-LLaMA-3.33-70B-v0.1
8
+ merge_method: sce
9
+ base_model: downloads/Llama-3.1-Nemotron-lorablated-70B
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  parameters:
11
+ select_topk: 0.17
 
12
  out_dtype: bfloat16
13
  tokenizer:
14
+ source: downloads/Negative_LLAMA_70B
model-00001-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ac1e861eee22ad067b8e9ef271f3a73ce1efa00dd783a5e0f72ff0828d453eb
3
  size 4672471504
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b3a24d56a6ded8e5622508a8fe2e8d1c7751648b255f6b8ebba7e0506a4a691
3
  size 4672471504
model-00002-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b4a1f5cbc31fe6f2dc2f4de1df0208a587ba8247b22f976bc7c673e0a12729b
3
  size 4664167400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cb14c55afd605d10e9a7505f56ceccc1023130d83130f31f5efbfbc78b4fe6d
3
  size 4664167400
model-00003-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a4ad906b79928056579d7c0d65914fafeac6c1639a87c2582855edf08fe0ca2
3
  size 4982917904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69d4be5b9c2070205eafef536061086979265c122782d43eb488e4184b227ee5
3
  size 4982917904
model-00004-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c9d37aacf099f7ae263196828ec2d2fe2f0ec9c70624b3e06fabf3188294df7
3
  size 4999711736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edc4540a70d03e1125e041cf4b462a3666d6b819cb60825f4cfe41576e1c8fd2
3
  size 4999711736
model-00005-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7f7ff20ad03503ab6c21b7f851bcb4f52203ed0a02d80565c29eec448a2e6f5
3
  size 4647373576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98def1d4e28e082c4351fed82b49bed660ed9257e5d7a9043b4fd53665fdf44a
3
  size 4647373576
model-00006-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f3e77fcb082b2d9019aed5661c47761f726a3c533249accfdb2a093dd718e932
3
  size 4664167400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:123283d4ca8bf6cdb5f2195075e94a0bab880128bddc4a8b13a66257e8c51b25
3
  size 4664167400
model-00007-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8562be87a33e2842f3ae2eb43e9331cf856292d68db78f3135f64ed17f2f7c50
3
  size 4664167416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8b3c6c161e92f18a0331f627ba0132d3db2eec128a727ac490831cd5392ac00
3
  size 4664167416
model-00008-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7f3b451fb4af113b9fe245e1850337eee6b4c7e470ad393899687c87708a5b2
3
  size 4982917904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:150bca9feb0bafd6c069b1acfe38c01bf9478b3e0b18d084a4da9a6a4b1793b5
3
  size 4982917904
model-00009-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b62b902ddd365fd70ab5f1d01073c439b610b24d3cff83bfcdf04431e213622
3
  size 4999711736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4a18a971f15aabcb861631d1587597d828e85df0edb1cd3017b909337973a94
3
  size 4999711736
model-00010-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab800f7aca9f67d2c41adf2c40ae0ac3b2c2dafc0bc9463a791fe6862f48d7bf
3
  size 4647373568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d522c71a859fee12c6ec16dea102efb2cd2c0b94ea591cd2df574ab1930ee49
3
  size 4647373568
model-00011-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d3cee0ba1914276363233150d5aadd6ebeadf43e6642c83bf7c9018504c73cc
3
  size 4664167408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db9501b975ba5e5fd9a5992d779b93c2bce864201c4ef69a5018a9e92557d746
3
  size 4664167408
model-00012-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ad5c369e8f2bca841d83173c6bcbd9c74f1612a08181c970b316906e3a36a9f
3
  size 4664167416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c53582030228400a0748f56a40723c0700f717adc52b1aa5076d3da535f138bf
3
  size 4664167416
model-00013-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95e3f8ca36f79351ce58ae07cd0b2f5c7fb53657de76404eeb5676b984ea9ba6
3
  size 4982917904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00deb9a49e9443d839b04afc952581f2b72b26d0442760d4615c6395f8f46916
3
  size 4982917904
model-00014-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab052080f454ab21db051b24140739bed3a15e0d1541941b36919ec407cc89c5
3
  size 4999711728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:722e7c7ea562f03ab62e51a44410009399accf7803c8ee7f0e7ea4ad0e49afb2
3
  size 4999711728
model-00015-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84e245b8b607debca85b49f2476929b9d2d27271af7d3fc8da4061a012f0d8a5
3
  size 4647373576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdb2a8bc013e16cd9b940e1e6664dcccd235c36a472b413b55e15a7c75432c42
3
  size 4647373576
model-00016-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5e6d71f9aff6e25c1c39f2dbdedbece9361e8238f3a7874e0822321d6f79ea6
3
  size 4664167408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6382b45299fb95de89a50adcaeddaf10937f2248ba4509683e0d2516eb59a645
3
  size 4664167408
model-00017-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63ee5afbde7ab2fa5e6457d04cc41a3a40fb7bef082a2344e6c976182dfb49e1
3
  size 4664167416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7091f5bfc1bc5a238307baf5409c5d7ad69bf9176f3ca0df65fc484381776d8
3
  size 4664167416
model-00018-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:babc6dcd3e2aa2781503320ebd4ebc121c179bcbcb9588c110a80033ba3b34b7
3
  size 4982917896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3aedf5b1fe34563ded1af4f560936d866eb4abf5ef8896918c480b62630a820
3
  size 4982917896
model-00019-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8d638afe0b07f538e049a150c48c0b45e918d3bfb1b4b3d01949544deb2088d
3
  size 4999711736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07ca7522dfd5e0a736b896cfb94333601b8516be3412630930aa29f282ed27a9
3
  size 4999711736
model-00020-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2fa78a673ad7d8bff0b7de91c1e7b7915a01cfb73dcbe23cbff6711e68824b5
3
  size 4647373576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a882ede88c3306e11d9f460174ef26726f2eec76fee439241ace8ddd0a81a9a0
3
  size 4647373576
model-00021-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca870f8c299f54cb4aa82579d49cd341346a5b6ef42a0f227b7fa80f0824655e
3
  size 4664167408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39f83347ff691ac4374b3035f1e92a4c23d015e57be86b9ee1b36cefcea519ea
3
  size 4664167408
model-00022-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f0e76539088af8995ba448c2d9be29f2b41741569361019af7431c5a417cc4d
3
  size 4664167408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:376ce963ef6e2cd92fdefc8e7cfcafdc7fb6aeb63cec7aad7fcf624797c438cf
3
  size 4664167408
model-00023-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c1c0ac33ee45ba69e4fdfd439c0ee8be9579172c2ebf641a9fa92bf301befad
3
  size 4982917904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05a2fd46a911832dfcc4fa7232b088d422687784c548a6f80d1a42d60d5044aa
3
  size 4982917904
model-00024-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf1bb7df0c51d50260433915530245d40a90222ac939182dc6874bd318cc83a5
3
  size 4999711736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77d2303a427bc2d8e406d0c516bb0f00025be0e645056d32b139b00160f1776e
3
  size 4999711736
model-00025-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d917173f0278ac11f0b0b90761795e157bada7bd5d31789566912e4edb817f20
3
  size 4647373576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c88da8dd2cde7156873641ae955ef30dc7e77b36f1fd8e73a95d419fd6f58317
3
  size 4647373576
model-00026-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3132caf9d8369c916532b8114a62b6b5d04a16937585d6460b61d0fd73de8fb0
3
  size 4664167400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca60fd1ab8777dc9f370601efb2e4f88487eb268a1d66d52547b6f95abc97e2f
3
  size 4664167400
model-00027-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc80d5bf4cfcbeafc251a56984695f4496ffd4d8cc6bf4b18a358fae2a587331
3
  size 4664167416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af369f3c7fd67fa6fd281b9068a5bcfc91c50b7f37f34e93a6678e1bda02ea07
3
  size 4664167416
model-00028-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8f68bcef76979d5e916dd46d234fd4af9d1a5dde2449e52268a5ffe31ad77a6
3
  size 4982917904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e53d961ddbaa0ad2b548105dd2982275e544d29c673e7cea0407ea9c16dfb66
3
  size 4982917904
model-00029-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3cb35f562ed9723fcc84d04fe0825c92a22309e208330f879be017214176def3
3
  size 4999711728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c6b0a0b89fe1fa4645e029d96fd0b8414b1c857c888dec204a081a450e03423
3
  size 4999711728
model-00030-of-00030.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4581b932440229bbe94b36193dc7d9d5b27d87264f26d4afe445019c797526e4
3
  size 1996539368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d16d4db6669ce7691891620001cdc0ff0941ca43916d93a306e00d16ff868b27
3
  size 1996539368
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json CHANGED
@@ -12,5 +12,12 @@
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
 
 
 
 
 
 
 
15
  }
16
  }
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|finetune_right_pad_id|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
  }
23
  }
tokenizer_config.json CHANGED
@@ -2050,7 +2050,7 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
- "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
  "extra_special_tokens": {},
@@ -2059,5 +2059,6 @@
2059
  "attention_mask"
2060
  ],
2061
  "model_max_length": 131072,
2062
- "tokenizer_class": "PreTrainedTokenizerFast"
 
2063
  }
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
  "extra_special_tokens": {},
 
2059
  "attention_mask"
2060
  ],
2061
  "model_max_length": 131072,
2062
+ "pad_token": "<|finetune_right_pad_id|>",
2063
+ "tokenizer_class": "PreTrainedTokenizer"
2064
  }