Delete duplicate eval
Browse files
README.md
CHANGED
|
@@ -175,18 +175,6 @@ pipeline = transformers.pipeline(
|
|
| 175 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
| 176 |
print(outputs[0]["generated_text"])
|
| 177 |
```
|
| 178 |
-
## [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
| 179 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Locutusque__OpenHercules-2.5-Mistral-7B)
|
| 180 |
-
|
| 181 |
-
| Metric |Value|
|
| 182 |
-
|---------------------------------|----:|
|
| 183 |
-
|Avg. |66.55|
|
| 184 |
-
|AI2 Reasoning Challenge (25-Shot)|64.25|
|
| 185 |
-
|HellaSwag (10-Shot) |84.84|
|
| 186 |
-
|MMLU (5-Shot) |64.21|
|
| 187 |
-
|TruthfulQA (0-shot) |47.84|
|
| 188 |
-
|Winogrande (5-shot) |78.93|
|
| 189 |
-
|GSM8k (5-shot) |59.21|
|
| 190 |
|
| 191 |
## Quants
|
| 192 |
|
|
|
|
| 175 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
| 176 |
print(outputs[0]["generated_text"])
|
| 177 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 178 |
|
| 179 |
## Quants
|
| 180 |
|