Eldar Kurtic commited on
Commit
919a5c6
·
1 Parent(s): bb81080

add evals for bf16

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -22,7 +22,7 @@ tags:
22
  - **Model Developers:** Neural Magic
23
 
24
  Quantized version of [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
25
- It achieves an average score of <TODO> on the [OpenLLM](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) benchmark (version 1), whereas the unquantized model achieves <TODO>.
26
 
27
  ### Model Optimizations
28
 
@@ -172,12 +172,12 @@ lm_eval \
172
 
173
  | Metric | mistralai/Mixtral-8x7B-Instruct-v0.1 | neuralmagic/Mixtral-8x7B-Instruct-v0.1-FP8 |
174
  |-----------------------------------------|:---------------------------------:|:-------------------------------------------:|
175
- | ARC-Challenge (Acc-Norm, 25-shot) | <TODO> | <TODO> |
176
- | GSM8K (Strict-Match, 5-shot) | <TODO> | <TODO> |
177
- | HellaSwag (Acc-Norm, 10-shot) | <TODO> | <TODO> |
178
- | MMLU (Acc, 5-shot) | <TODO> | <TODO> |
179
- | TruthfulQA (MC2, 0-shot) | <TODO> | <TODO> |
180
- | Winogrande (Acc, 5-shot) | <TODO> | <TODO> |
181
- | **Average Score** | **<TODO>** | **<TODO>** |
182
  | **Recovery (%)** | **100.00** | **<TODO>** |
183
 
 
22
  - **Model Developers:** Neural Magic
23
 
24
  Quantized version of [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1).
25
+ It achieves an average score of <TODO> on the [OpenLLM](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) benchmark (version 1), whereas the unquantized model achieves 73.44.
26
 
27
  ### Model Optimizations
28
 
 
172
 
173
  | Metric | mistralai/Mixtral-8x7B-Instruct-v0.1 | neuralmagic/Mixtral-8x7B-Instruct-v0.1-FP8 |
174
  |-----------------------------------------|:---------------------------------:|:-------------------------------------------:|
175
+ | ARC-Challenge (Acc-Norm, 25-shot) | 70.48 | <TODO> |
176
+ | GSM8K (Strict-Match, 5-shot) | 65.50 | <TODO> |
177
+ | HellaSwag (Acc-Norm, 10-shot) | 87.33 | <TODO> |
178
+ | MMLU (Acc, 5-shot) | 70.30 | <TODO> |
179
+ | TruthfulQA (MC2, 0-shot) | 64.81 | <TODO> |
180
+ | Winogrande (Acc, 5-shot) | 82.24 | <TODO> |
181
+ | **Average Score** | **73.44** | **<TODO>** |
182
  | **Recovery (%)** | **100.00** | **<TODO>** |
183