bartowski commited on
Commit
ae1ffb1
·
verified ·
1 Parent(s): 4376d9b

Llamacpp quants

Browse files
.gitattributes CHANGED
@@ -33,3 +33,16 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Hercules-3.1-Mistral-7B-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Hercules-3.1-Mistral-7B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Hercules-3.1-Mistral-7B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Hercules-3.1-Mistral-7B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Hercules-3.1-Mistral-7B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Hercules-3.1-Mistral-7B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Hercules-3.1-Mistral-7B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Hercules-3.1-Mistral-7B-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Hercules-3.1-Mistral-7B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Hercules-3.1-Mistral-7B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Hercules-3.1-Mistral-7B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Hercules-3.1-Mistral-7B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Hercules-3.1-Mistral-7B-fp16.gguf filter=lfs diff=lfs merge=lfs -text
Hercules-3.1-Mistral-7B-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8597174984cb9e2f7f914c9ee767cbc6c248703222ba8fa4bb1a9b23a8ccac4a
3
+ size 2719241984
Hercules-3.1-Mistral-7B-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:512f541d3886a7d3a661ef3bd044ac7b74c168c28ab22ee5a4b44d4f68f17aa4
3
+ size 3822024448
Hercules-3.1-Mistral-7B-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9df8d1cab8ec62358be6314d6ab881e10de3cc807d688d127b6e747574e831df
3
+ size 3518985984
Hercules-3.1-Mistral-7B-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bf20da6761353fb4514f24cf051cbe5c20d3c860f7a31eaef1e0ad9835a88d8
3
+ size 3164567296
Hercules-3.1-Mistral-7B-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745629bf898b12b963cbccfc2c7c8e4d849171782e077e88115a1d8758560c87
3
+ size 4108916480
Hercules-3.1-Mistral-7B-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e6645d32437f9765ee350e808ec0acb58499a76cb2454d7de6fbc38f34968c9
3
+ size 4368439040
Hercules-3.1-Mistral-7B-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83ba3ee3a3a2f06d6503f4d1d8ac0e3e5b9e936d68f7f958a9cde1d0f2c27f1e
3
+ size 4140373760
Hercules-3.1-Mistral-7B-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a4a13c86c5389881b55116a57181e815749b2408f9eafe3d15f8948f21c2fe2
3
+ size 4997715712
Hercules-3.1-Mistral-7B-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6df84d2cf1d8c8a783163d87ab6199e0669946de40d04b195d2a5ebf1798245
3
+ size 5131409152
Hercules-3.1-Mistral-7B-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf665a4af9f745b2999318653691ab310652bdcb9162198417bcfddbfd20f0b2
3
+ size 4997715712
Hercules-3.1-Mistral-7B-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdf618abf854cb0e702f9f6c62c8033a89d0a0457033471da3b9ab4ce677216f
3
+ size 5942064896
Hercules-3.1-Mistral-7B-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:829b60d1fadde45c7389b07f096cc8419a9e413595781938341fb4ce96ab4561
3
+ size 7695857408
Hercules-3.1-Mistral-7B-fp16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4529fbd18f86c37abf59e5749a7e3617f7e2809dd68f34210969bf506bc05e2d
3
+ size 14484731584
README.md ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ tags:
5
+ - chemistry
6
+ - biology
7
+ - code
8
+ - medical
9
+ - not-for-all-audiences
10
+ datasets:
11
+ - Locutusque/Hercules-v3.0
12
+ model-index:
13
+ - name: Hercules-3.1-Mistral-7B
14
+ results:
15
+ - task:
16
+ type: text-generation
17
+ name: Text Generation
18
+ dataset:
19
+ name: AI2 Reasoning Challenge (25-Shot)
20
+ type: ai2_arc
21
+ config: ARC-Challenge
22
+ split: test
23
+ args:
24
+ num_few_shot: 25
25
+ metrics:
26
+ - type: acc_norm
27
+ value: 61.18
28
+ name: normalized accuracy
29
+ source:
30
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/Hercules-3.1-Mistral-7B
31
+ name: Open LLM Leaderboard
32
+ - task:
33
+ type: text-generation
34
+ name: Text Generation
35
+ dataset:
36
+ name: HellaSwag (10-Shot)
37
+ type: hellaswag
38
+ split: validation
39
+ args:
40
+ num_few_shot: 10
41
+ metrics:
42
+ - type: acc_norm
43
+ value: 83.55
44
+ name: normalized accuracy
45
+ source:
46
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/Hercules-3.1-Mistral-7B
47
+ name: Open LLM Leaderboard
48
+ - task:
49
+ type: text-generation
50
+ name: Text Generation
51
+ dataset:
52
+ name: MMLU (5-Shot)
53
+ type: cais/mmlu
54
+ config: all
55
+ split: test
56
+ args:
57
+ num_few_shot: 5
58
+ metrics:
59
+ - type: acc
60
+ value: 63.65
61
+ name: accuracy
62
+ source:
63
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/Hercules-3.1-Mistral-7B
64
+ name: Open LLM Leaderboard
65
+ - task:
66
+ type: text-generation
67
+ name: Text Generation
68
+ dataset:
69
+ name: TruthfulQA (0-shot)
70
+ type: truthful_qa
71
+ config: multiple_choice
72
+ split: validation
73
+ args:
74
+ num_few_shot: 0
75
+ metrics:
76
+ - type: mc2
77
+ value: 42.83
78
+ source:
79
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/Hercules-3.1-Mistral-7B
80
+ name: Open LLM Leaderboard
81
+ - task:
82
+ type: text-generation
83
+ name: Text Generation
84
+ dataset:
85
+ name: Winogrande (5-shot)
86
+ type: winogrande
87
+ config: winogrande_xl
88
+ split: validation
89
+ args:
90
+ num_few_shot: 5
91
+ metrics:
92
+ - type: acc
93
+ value: 79.01
94
+ name: accuracy
95
+ source:
96
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/Hercules-3.1-Mistral-7B
97
+ name: Open LLM Leaderboard
98
+ - task:
99
+ type: text-generation
100
+ name: Text Generation
101
+ dataset:
102
+ name: GSM8k (5-shot)
103
+ type: gsm8k
104
+ config: main
105
+ split: test
106
+ args:
107
+ num_few_shot: 5
108
+ metrics:
109
+ - type: acc
110
+ value: 42.3
111
+ name: accuracy
112
+ source:
113
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Locutusque/Hercules-3.1-Mistral-7B
114
+ name: Open LLM Leaderboard
115
+ quantized_by: bartowski
116
+ pipeline_tag: text-generation
117
+ ---
118
+
119
+ ## Llamacpp Quantizations of Hercules-3.1-Mistral-7B
120
+
121
+ Using <a href="https://github.com/ggerganov/llama.cpp/commit/fa974646e1a2024fc7dc9e6f27cf1f2f5d4a3763">llama.cpp commit fa97464</a> for quantization.
122
+
123
+ Original model: https://huggingface.co/Locutusque/Hercules-3.1-Mistral-7B
124
+
125
+ Download a file (not the whole branch) from below:
126
+
127
+ | Filename | Quant type | File Size | Description |
128
+ | -------- | ---------- | --------- | ----------- |
129
+ | [Hercules-3.1-Mistral-7B-Q8_0.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q8_0.gguf) | Q8_0 | 7.69GB | Extremely high quality, generally unneeded but max available quant. |
130
+ | [Hercules-3.1-Mistral-7B-Q6_K.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q6_K.gguf) | Q6_K | 5.94GB | Very high quality, near perfect, *recommended*. |
131
+ | [Hercules-3.1-Mistral-7B-Q5_K_M.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q5_K_M.gguf) | Q5_K_M | 5.13GB | High quality, very usable. |
132
+ | [Hercules-3.1-Mistral-7B-Q5_K_S.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q5_K_S.gguf) | Q5_K_S | 4.99GB | High quality, very usable. |
133
+ | [Hercules-3.1-Mistral-7B-Q5_0.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q5_0.gguf) | Q5_0 | 4.99GB | High quality, older format, generally not recommended. |
134
+ | [Hercules-3.1-Mistral-7B-Q4_K_M.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q4_K_M.gguf) | Q4_K_M | 4.36GB | Good quality, similar to 4.25 bpw. |
135
+ | [Hercules-3.1-Mistral-7B-Q4_K_S.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q4_K_S.gguf) | Q4_K_S | 4.14GB | Slightly lower quality with small space savings. |
136
+ | [Hercules-3.1-Mistral-7B-Q4_0.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q4_0.gguf) | Q4_0 | 4.10GB | Decent quality, older format, generally not recommended. |
137
+ | [Hercules-3.1-Mistral-7B-Q3_K_L.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q3_K_L.gguf) | Q3_K_L | 3.82GB | Lower quality but usable, good for low RAM availability. |
138
+ | [Hercules-3.1-Mistral-7B-Q3_K_M.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q3_K_M.gguf) | Q3_K_M | 3.51GB | Even lower quality. |
139
+ | [Hercules-3.1-Mistral-7B-Q3_K_S.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q3_K_S.gguf) | Q3_K_S | 3.16GB | Low quality, not recommended. |
140
+ | [Hercules-3.1-Mistral-7B-Q2_K.gguf](https://huggingface.co/bartowski/Hercules-3.1-Mistral-7B-GGUF/blob/main/Hercules-3.1-Mistral-7B-Q2_K.gguf) | Q2_K | 2.71GB | Extremely low quality, *not* recommended.
141
+
142
+ Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski