Eldar Kurtic
commited on
Commit
·
919a5c6
1
Parent(s):
bb81080
add evals for bf16
Browse files
README.md
CHANGED
@@ -22,7 +22,7 @@ tags:
|
|
22 |
- **Model Developers:** Neural Magic
|
23 |
|
24 |
Quantized version of [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
|
25 |
-
It achieves an average score of <TODO> on the [OpenLLM](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) benchmark (version 1), whereas the unquantized model achieves
|
26 |
|
27 |
### Model Optimizations
|
28 |
|
@@ -172,12 +172,12 @@ lm_eval \
|
|
172 |
|
173 |
| Metric | mistralai/Mixtral-8x7B-Instruct-v0.1 | neuralmagic/Mixtral-8x7B-Instruct-v0.1-FP8 |
|
174 |
|-----------------------------------------|:---------------------------------:|:-------------------------------------------:|
|
175 |
-
| ARC-Challenge (Acc-Norm, 25-shot) |
|
176 |
-
| GSM8K (Strict-Match, 5-shot) |
|
177 |
-
| HellaSwag (Acc-Norm, 10-shot) |
|
178 |
-
| MMLU (Acc, 5-shot) |
|
179 |
-
| TruthfulQA (MC2, 0-shot) |
|
180 |
-
| Winogrande (Acc, 5-shot) |
|
181 |
-
| **Average Score** |
|
182 |
| **Recovery (%)** | **100.00** | **<TODO>** |
|
183 |
|
|
|
22 |
- **Model Developers:** Neural Magic
|
23 |
|
24 |
Quantized version of [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
|
25 |
+
It achieves an average score of <TODO> on the [OpenLLM](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) benchmark (version 1), whereas the unquantized model achieves 73.44.
|
26 |
|
27 |
### Model Optimizations
|
28 |
|
|
|
172 |
|
173 |
| Metric | mistralai/Mixtral-8x7B-Instruct-v0.1 | neuralmagic/Mixtral-8x7B-Instruct-v0.1-FP8 |
|
174 |
|-----------------------------------------|:---------------------------------:|:-------------------------------------------:|
|
175 |
+
| ARC-Challenge (Acc-Norm, 25-shot) | 70.48 | <TODO> |
|
176 |
+
| GSM8K (Strict-Match, 5-shot) | 65.50 | <TODO> |
|
177 |
+
| HellaSwag (Acc-Norm, 10-shot) | 87.33 | <TODO> |
|
178 |
+
| MMLU (Acc, 5-shot) | 70.30 | <TODO> |
|
179 |
+
| TruthfulQA (MC2, 0-shot) | 64.81 | <TODO> |
|
180 |
+
| Winogrande (Acc, 5-shot) | 82.24 | <TODO> |
|
181 |
+
| **Average Score** | **73.44** | **<TODO>** |
|
182 |
| **Recovery (%)** | **100.00** | **<TODO>** |
|
183 |
|