Update README.md
Browse files
README.md
CHANGED
@@ -3,6 +3,8 @@ license: gpl-3.0
|
|
3 |
language:
|
4 |
- en
|
5 |
pipeline_tag: text-generation
|
|
|
|
|
6 |
---
|
7 |
|
8 |
# Phi-4-instruct-abliterated
|
@@ -138,4 +140,4 @@ At the high-level overview of the model quality on representative benchmarks. Fo
|
|
138 |
| Factual Knowledge | SimpleQA | 3.0 | 7.6 | 5.4 | 9.9 | 20.9 | 10.2 | **39.4** |
|
139 |
| Reasoning | DROP | 75.5 | 68.3 | 85.5 | 79.3 | **90.2** | 76.7 | 80.9 |
|
140 |
|
141 |
-
\* These scores are lower than those reported by Meta, perhaps because simple-evals has a strict formatting requirement that Llama models have particular trouble following. We use the simple-evals framework because it is reproducible, but Meta reports 77 for MATH and 88 for HumanEval on Llama-3.3-70B.
|
|
|
3 |
language:
|
4 |
- en
|
5 |
pipeline_tag: text-generation
|
6 |
+
base_model:
|
7 |
+
- NyxKrage/Microsoft_Phi-4
|
8 |
---
|
9 |
|
10 |
# Phi-4-instruct-abliterated
|
|
|
140 |
| Factual Knowledge | SimpleQA | 3.0 | 7.6 | 5.4 | 9.9 | 20.9 | 10.2 | **39.4** |
|
141 |
| Reasoning | DROP | 75.5 | 68.3 | 85.5 | 79.3 | **90.2** | 76.7 | 80.9 |
|
142 |
|
143 |
+
\* These scores are lower than those reported by Meta, perhaps because simple-evals has a strict formatting requirement that Llama models have particular trouble following. We use the simple-evals framework because it is reproducible, but Meta reports 77 for MATH and 88 for HumanEval on Llama-3.3-70B.
|