Add files using upload-large-folder tool
Browse files- .ipynb_checkpoints/mergekit_config-checkpoint.yml +14 -0
- README.md +19 -40
- config.json +2 -2
- mergekit_config.yml +10 -25
- model-00001-of-00030.safetensors +1 -1
- model-00002-of-00030.safetensors +1 -1
- model-00003-of-00030.safetensors +1 -1
- model-00004-of-00030.safetensors +1 -1
- model-00005-of-00030.safetensors +1 -1
- model-00006-of-00030.safetensors +1 -1
- model-00007-of-00030.safetensors +1 -1
- model-00008-of-00030.safetensors +1 -1
- model-00009-of-00030.safetensors +1 -1
- model-00010-of-00030.safetensors +1 -1
- model-00011-of-00030.safetensors +1 -1
- model-00012-of-00030.safetensors +1 -1
- model-00013-of-00030.safetensors +1 -1
- model-00014-of-00030.safetensors +1 -1
- model-00015-of-00030.safetensors +1 -1
- model-00016-of-00030.safetensors +1 -1
- model-00017-of-00030.safetensors +1 -1
- model-00018-of-00030.safetensors +1 -1
- model-00019-of-00030.safetensors +1 -1
- model-00020-of-00030.safetensors +1 -1
- model-00021-of-00030.safetensors +1 -1
- model-00022-of-00030.safetensors +1 -1
- model-00023-of-00030.safetensors +1 -1
- model-00024-of-00030.safetensors +1 -1
- model-00025-of-00030.safetensors +1 -1
- model-00026-of-00030.safetensors +1 -1
- model-00027-of-00030.safetensors +1 -1
- model-00028-of-00030.safetensors +1 -1
- model-00029-of-00030.safetensors +1 -1
- model-00030-of-00030.safetensors +1 -1
- model.safetensors.index.json +0 -0
- special_tokens_map.json +7 -0
- tokenizer_config.json +3 -2
.ipynb_checkpoints/mergekit_config-checkpoint.yml
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
models:
|
2 |
+
- model: downloads/L3.1-70B-Hanami-x1
|
3 |
+
- model: downloads/70B-L3.3-Cirrus-x1
|
4 |
+
- model: downloads/Wayfarer-Large-70B-Llama-3.3
|
5 |
+
- model: downloads/Negative_LLAMA_70B
|
6 |
+
- model: downloads/Anubis-70B-v1
|
7 |
+
- model: downloads/EVA-LLaMA-3.33-70B-v0.1
|
8 |
+
merge_method: sce
|
9 |
+
base_model: downloads/Llama-3.1-Nemotron-lorablated-70B
|
10 |
+
parameters:
|
11 |
+
select_topk: 0.17
|
12 |
+
out_dtype: bfloat16
|
13 |
+
tokenizer:
|
14 |
+
source: downloads/Negative_LLAMA_70B
|
README.md
CHANGED
@@ -1,34 +1,29 @@
|
|
1 |
---
|
2 |
-
base_model:
|
3 |
-
- EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
|
4 |
-
- nbeerbower/Llama-3.1-Nemotron-lorablated-70B
|
5 |
-
- Sao10K/L3.1-70B-Hanami-x1
|
6 |
-
- SicariusSicariiStuff/Negative_LLAMA_70B
|
7 |
-
- TheDrummer/Anubis-70B-v1
|
8 |
-
- LatitudeGames/Wayfarer-Large-70B-Llama-3.3
|
9 |
library_name: transformers
|
10 |
tags:
|
11 |
- mergekit
|
12 |
- merge
|
13 |
|
14 |
---
|
15 |
-
#
|
16 |
|
17 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
18 |
|
19 |
## Merge Details
|
20 |
### Merge Method
|
21 |
|
22 |
-
This model was merged using the [
|
23 |
|
24 |
### Models Merged
|
25 |
|
26 |
The following models were included in the merge:
|
27 |
-
*
|
28 |
-
*
|
29 |
-
*
|
30 |
-
*
|
31 |
-
*
|
|
|
32 |
|
33 |
### Configuration
|
34 |
|
@@ -36,33 +31,17 @@ The following YAML configuration was used to produce this model:
|
|
36 |
|
37 |
```yaml
|
38 |
models:
|
39 |
-
- model:
|
40 |
-
|
41 |
-
|
42 |
-
|
43 |
-
- model:
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
- model: SicariusSicariiStuff/Negative_LLAMA_70B
|
48 |
-
parameters:
|
49 |
-
weight: 0.20
|
50 |
-
density: 0.7
|
51 |
-
- model: TheDrummer/Anubis-70B-v1
|
52 |
-
parameters:
|
53 |
-
weight: 0.20
|
54 |
-
density: 0.7
|
55 |
-
- model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
|
56 |
-
parameters:
|
57 |
-
weight: 0.20
|
58 |
-
density: 0.7
|
59 |
-
merge_method: della_linear
|
60 |
-
base_model: nbeerbower/Llama-3.1-Nemotron-lorablated-70B
|
61 |
parameters:
|
62 |
-
|
63 |
-
lambda: 1.1
|
64 |
out_dtype: bfloat16
|
65 |
tokenizer:
|
66 |
-
source:
|
67 |
-
|
68 |
```
|
|
|
1 |
---
|
2 |
+
base_model: []
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
library_name: transformers
|
4 |
tags:
|
5 |
- mergekit
|
6 |
- merge
|
7 |
|
8 |
---
|
9 |
+
# PrimogentorSCE3
|
10 |
|
11 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
12 |
|
13 |
## Merge Details
|
14 |
### Merge Method
|
15 |
|
16 |
+
This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using downloads/Llama-3.1-Nemotron-lorablated-70B as a base.
|
17 |
|
18 |
### Models Merged
|
19 |
|
20 |
The following models were included in the merge:
|
21 |
+
* downloads/Wayfarer-Large-70B-Llama-3.3
|
22 |
+
* downloads/EVA-LLaMA-3.33-70B-v0.1
|
23 |
+
* downloads/Negative_LLAMA_70B
|
24 |
+
* downloads/L3.1-70B-Hanami-x1
|
25 |
+
* downloads/70B-L3.3-Cirrus-x1
|
26 |
+
* downloads/Anubis-70B-v1
|
27 |
|
28 |
### Configuration
|
29 |
|
|
|
31 |
|
32 |
```yaml
|
33 |
models:
|
34 |
+
- model: downloads/L3.1-70B-Hanami-x1
|
35 |
+
- model: downloads/70B-L3.3-Cirrus-x1
|
36 |
+
- model: downloads/Wayfarer-Large-70B-Llama-3.3
|
37 |
+
- model: downloads/Negative_LLAMA_70B
|
38 |
+
- model: downloads/Anubis-70B-v1
|
39 |
+
- model: downloads/EVA-LLaMA-3.33-70B-v0.1
|
40 |
+
merge_method: sce
|
41 |
+
base_model: downloads/Llama-3.1-Nemotron-lorablated-70B
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
42 |
parameters:
|
43 |
+
select_topk: 0.17
|
|
|
44 |
out_dtype: bfloat16
|
45 |
tokenizer:
|
46 |
+
source: downloads/Negative_LLAMA_70B
|
|
|
47 |
```
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
@@ -34,7 +34,7 @@
|
|
34 |
"rope_theta": 500000.0,
|
35 |
"tie_word_embeddings": false,
|
36 |
"torch_dtype": "bfloat16",
|
37 |
-
"transformers_version": "4.
|
38 |
"use_cache": true,
|
39 |
"vocab_size": 128256
|
40 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "downloads/Llama-3.1-Nemotron-lorablated-70B",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
34 |
"rope_theta": 500000.0,
|
35 |
"tie_word_embeddings": false,
|
36 |
"torch_dtype": "bfloat16",
|
37 |
+
"transformers_version": "4.49.0",
|
38 |
"use_cache": true,
|
39 |
"vocab_size": 128256
|
40 |
}
|
mergekit_config.yml
CHANGED
@@ -1,29 +1,14 @@
|
|
1 |
models:
|
2 |
-
- model:
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
- model:
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
- model: SicariusSicariiStuff/Negative_LLAMA_70B
|
11 |
-
parameters:
|
12 |
-
weight: 0.20
|
13 |
-
density: 0.7
|
14 |
-
- model: TheDrummer/Anubis-70B-v1
|
15 |
-
parameters:
|
16 |
-
weight: 0.20
|
17 |
-
density: 0.7
|
18 |
-
- model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
|
19 |
-
parameters:
|
20 |
-
weight: 0.20
|
21 |
-
density: 0.7
|
22 |
-
merge_method: della_linear
|
23 |
-
base_model: nbeerbower/Llama-3.1-Nemotron-lorablated-70B
|
24 |
parameters:
|
25 |
-
|
26 |
-
lambda: 1.1
|
27 |
out_dtype: bfloat16
|
28 |
tokenizer:
|
29 |
-
source:
|
|
|
1 |
models:
|
2 |
+
- model: downloads/L3.1-70B-Hanami-x1
|
3 |
+
- model: downloads/70B-L3.3-Cirrus-x1
|
4 |
+
- model: downloads/Wayfarer-Large-70B-Llama-3.3
|
5 |
+
- model: downloads/Negative_LLAMA_70B
|
6 |
+
- model: downloads/Anubis-70B-v1
|
7 |
+
- model: downloads/EVA-LLaMA-3.33-70B-v0.1
|
8 |
+
merge_method: sce
|
9 |
+
base_model: downloads/Llama-3.1-Nemotron-lorablated-70B
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
parameters:
|
11 |
+
select_topk: 0.17
|
|
|
12 |
out_dtype: bfloat16
|
13 |
tokenizer:
|
14 |
+
source: downloads/Negative_LLAMA_70B
|
model-00001-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4672471504
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2b3a24d56a6ded8e5622508a8fe2e8d1c7751648b255f6b8ebba7e0506a4a691
|
3 |
size 4672471504
|
model-00002-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167400
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5cb14c55afd605d10e9a7505f56ceccc1023130d83130f31f5efbfbc78b4fe6d
|
3 |
size 4664167400
|
model-00003-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4982917904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69d4be5b9c2070205eafef536061086979265c122782d43eb488e4184b227ee5
|
3 |
size 4982917904
|
model-00004-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999711736
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edc4540a70d03e1125e041cf4b462a3666d6b819cb60825f4cfe41576e1c8fd2
|
3 |
size 4999711736
|
model-00005-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4647373576
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:98def1d4e28e082c4351fed82b49bed660ed9257e5d7a9043b4fd53665fdf44a
|
3 |
size 4647373576
|
model-00006-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167400
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:123283d4ca8bf6cdb5f2195075e94a0bab880128bddc4a8b13a66257e8c51b25
|
3 |
size 4664167400
|
model-00007-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167416
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8b3c6c161e92f18a0331f627ba0132d3db2eec128a727ac490831cd5392ac00
|
3 |
size 4664167416
|
model-00008-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4982917904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:150bca9feb0bafd6c069b1acfe38c01bf9478b3e0b18d084a4da9a6a4b1793b5
|
3 |
size 4982917904
|
model-00009-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999711736
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c4a18a971f15aabcb861631d1587597d828e85df0edb1cd3017b909337973a94
|
3 |
size 4999711736
|
model-00010-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4647373568
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d522c71a859fee12c6ec16dea102efb2cd2c0b94ea591cd2df574ab1930ee49
|
3 |
size 4647373568
|
model-00011-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167408
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db9501b975ba5e5fd9a5992d779b93c2bce864201c4ef69a5018a9e92557d746
|
3 |
size 4664167408
|
model-00012-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167416
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c53582030228400a0748f56a40723c0700f717adc52b1aa5076d3da535f138bf
|
3 |
size 4664167416
|
model-00013-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4982917904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:00deb9a49e9443d839b04afc952581f2b72b26d0442760d4615c6395f8f46916
|
3 |
size 4982917904
|
model-00014-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999711728
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:722e7c7ea562f03ab62e51a44410009399accf7803c8ee7f0e7ea4ad0e49afb2
|
3 |
size 4999711728
|
model-00015-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4647373576
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cdb2a8bc013e16cd9b940e1e6664dcccd235c36a472b413b55e15a7c75432c42
|
3 |
size 4647373576
|
model-00016-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167408
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6382b45299fb95de89a50adcaeddaf10937f2248ba4509683e0d2516eb59a645
|
3 |
size 4664167408
|
model-00017-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167416
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c7091f5bfc1bc5a238307baf5409c5d7ad69bf9176f3ca0df65fc484381776d8
|
3 |
size 4664167416
|
model-00018-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4982917896
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3aedf5b1fe34563ded1af4f560936d866eb4abf5ef8896918c480b62630a820
|
3 |
size 4982917896
|
model-00019-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999711736
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:07ca7522dfd5e0a736b896cfb94333601b8516be3412630930aa29f282ed27a9
|
3 |
size 4999711736
|
model-00020-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4647373576
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a882ede88c3306e11d9f460174ef26726f2eec76fee439241ace8ddd0a81a9a0
|
3 |
size 4647373576
|
model-00021-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167408
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39f83347ff691ac4374b3035f1e92a4c23d015e57be86b9ee1b36cefcea519ea
|
3 |
size 4664167408
|
model-00022-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167408
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:376ce963ef6e2cd92fdefc8e7cfcafdc7fb6aeb63cec7aad7fcf624797c438cf
|
3 |
size 4664167408
|
model-00023-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4982917904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:05a2fd46a911832dfcc4fa7232b088d422687784c548a6f80d1a42d60d5044aa
|
3 |
size 4982917904
|
model-00024-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999711736
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:77d2303a427bc2d8e406d0c516bb0f00025be0e645056d32b139b00160f1776e
|
3 |
size 4999711736
|
model-00025-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4647373576
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c88da8dd2cde7156873641ae955ef30dc7e77b36f1fd8e73a95d419fd6f58317
|
3 |
size 4647373576
|
model-00026-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167400
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca60fd1ab8777dc9f370601efb2e4f88487eb268a1d66d52547b6f95abc97e2f
|
3 |
size 4664167400
|
model-00027-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4664167416
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af369f3c7fd67fa6fd281b9068a5bcfc91c50b7f37f34e93a6678e1bda02ea07
|
3 |
size 4664167416
|
model-00028-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4982917904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e53d961ddbaa0ad2b548105dd2982275e544d29c673e7cea0407ea9c16dfb66
|
3 |
size 4982917904
|
model-00029-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999711728
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c6b0a0b89fe1fa4645e029d96fd0b8414b1c857c888dec204a081a450e03423
|
3 |
size 4999711728
|
model-00030-of-00030.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1996539368
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d16d4db6669ce7691891620001cdc0ff0941ca43916d93a306e00d16ff868b27
|
3 |
size 1996539368
|
model.safetensors.index.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
special_tokens_map.json
CHANGED
@@ -12,5 +12,12 @@
|
|
12 |
"normalized": false,
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
}
|
16 |
}
|
|
|
12 |
"normalized": false,
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "<|finetune_right_pad_id|>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
}
|
23 |
}
|
tokenizer_config.json
CHANGED
@@ -2050,7 +2050,7 @@
|
|
2050 |
}
|
2051 |
},
|
2052 |
"bos_token": "<|begin_of_text|>",
|
2053 |
-
"chat_template": "{
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
2056 |
"extra_special_tokens": {},
|
@@ -2059,5 +2059,6 @@
|
|
2059 |
"attention_mask"
|
2060 |
],
|
2061 |
"model_max_length": 131072,
|
2062 |
-
"
|
|
|
2063 |
}
|
|
|
2050 |
}
|
2051 |
},
|
2052 |
"bos_token": "<|begin_of_text|>",
|
2053 |
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
|
2054 |
"clean_up_tokenization_spaces": true,
|
2055 |
"eos_token": "<|eot_id|>",
|
2056 |
"extra_special_tokens": {},
|
|
|
2059 |
"attention_mask"
|
2060 |
],
|
2061 |
"model_max_length": 131072,
|
2062 |
+
"pad_token": "<|finetune_right_pad_id|>",
|
2063 |
+
"tokenizer_class": "PreTrainedTokenizer"
|
2064 |
}
|