Orion-zhen commited on
Commit
caaff26
·
verified ·
1 Parent(s): a0f8aa4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -3,6 +3,8 @@ license: gpl-3.0
3
  language:
4
  - en
5
  pipeline_tag: text-generation
 
 
6
  ---
7
 
8
  # Phi-4-instruct-abliterated
@@ -138,4 +140,4 @@ At the high-level overview of the model quality on representative benchmarks. Fo
138
  | Factual Knowledge | SimpleQA | 3.0 | 7.6 | 5.4 | 9.9 | 20.9 | 10.2 | **39.4** |
139
  | Reasoning | DROP | 75.5 | 68.3 | 85.5 | 79.3 | **90.2** | 76.7 | 80.9 |
140
 
141
- \* These scores are lower than those reported by Meta, perhaps because simple-evals has a strict formatting requirement that Llama models have particular trouble following. We use the simple-evals framework because it is reproducible, but Meta reports 77 for MATH and 88 for HumanEval on Llama-3.3-70B.
 
3
  language:
4
  - en
5
  pipeline_tag: text-generation
6
+ base_model:
7
+ - NyxKrage/Microsoft_Phi-4
8
  ---
9
 
10
  # Phi-4-instruct-abliterated
 
140
  | Factual Knowledge | SimpleQA | 3.0 | 7.6 | 5.4 | 9.9 | 20.9 | 10.2 | **39.4** |
141
  | Reasoning | DROP | 75.5 | 68.3 | 85.5 | 79.3 | **90.2** | 76.7 | 80.9 |
142
 
143
+ \* These scores are lower than those reported by Meta, perhaps because simple-evals has a strict formatting requirement that Llama models have particular trouble following. We use the simple-evals framework because it is reproducible, but Meta reports 77 for MATH and 88 for HumanEval on Llama-3.3-70B.