Update README.md
Browse files
README.md
CHANGED
|
@@ -140,13 +140,13 @@ model-index:
|
|
| 140 |
<h2>Merged Evals (Has Not Been Finetuned):</h2>
|
| 141 |
<p>Aura-llama</p>
|
| 142 |
<ul>
|
| 143 |
-
<li>Avg:
|
| 144 |
-
<li>ARC:
|
| 145 |
-
<li>HellaSwag:
|
| 146 |
-
<li>MMLU:
|
| 147 |
-
<li>T-QA:
|
| 148 |
-
<li>Winogrande:
|
| 149 |
-
<li>GSM8K:
|
| 150 |
</ul>
|
| 151 |
</div>
|
| 152 |
<div class="update-section">
|
|
@@ -173,14 +173,3 @@ slices:
|
|
| 173 |
</body>
|
| 174 |
</html>
|
| 175 |
|
| 176 |
-
|
| 177 |
-
| Metric |Value|
|
| 178 |
-
|---------------------------------|----:|
|
| 179 |
-
|Avg. |63.13|
|
| 180 |
-
|AI2 Reasoning Challenge (25-Shot)|58.02|
|
| 181 |
-
|HellaSwag (10-Shot) |77.82|
|
| 182 |
-
|MMLU (5-Shot) |65.61|
|
| 183 |
-
|TruthfulQA (0-shot) |51.94|
|
| 184 |
-
|Winogrande (5-shot) |73.40|
|
| 185 |
-
|GSM8k (5-shot) |52.01|
|
| 186 |
-
|
|
|
|
| 140 |
<h2>Merged Evals (Has Not Been Finetuned):</h2>
|
| 141 |
<p>Aura-llama</p>
|
| 142 |
<ul>
|
| 143 |
+
<li>Avg: 63.13</li>
|
| 144 |
+
<li>ARC: 58.02</li>
|
| 145 |
+
<li>HellaSwag: 77.82</li>
|
| 146 |
+
<li>MMLU: 65.61</li>
|
| 147 |
+
<li>T-QA: 51.94</li>
|
| 148 |
+
<li>Winogrande: 73.40</li>
|
| 149 |
+
<li>GSM8K: 52.01</li>
|
| 150 |
</ul>
|
| 151 |
</div>
|
| 152 |
<div class="update-section">
|
|
|
|
| 173 |
</body>
|
| 174 |
</html>
|
| 175 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|