Readme update
Browse files
README.md
CHANGED
|
@@ -16,7 +16,7 @@ pinned: false
|
|
| 16 |
The traditional evaluation of NLP labeled spans with precision, recall, and F1-score leads to double penalties for
|
| 17 |
close-to-correct annotations. As [Manning (2006)](https://nlpers.blogspot.com/2006/08/doing-named-entity-recognition-dont.html)
|
| 18 |
argues in an article about named entity recognition, this can lead to undesirable effects when systems are optimized for these traditional metrics.
|
| 19 |
-
|
| 20 |
|
| 21 |
## How to Use
|
| 22 |
FairEval outputs the error count (TP, FP, etc.) and resulting scores (Precision, Recall and F1) from a reference list of
|
|
|
|
| 16 |
The traditional evaluation of NLP labeled spans with precision, recall, and F1-score leads to double penalties for
|
| 17 |
close-to-correct annotations. As [Manning (2006)](https://nlpers.blogspot.com/2006/08/doing-named-entity-recognition-dont.html)
|
| 18 |
argues in an article about named entity recognition, this can lead to undesirable effects when systems are optimized for these traditional metrics.
|
| 19 |
+
To address these issues, this metric provides an implementation of FairEval, proposed by [Ortmann (2022)](https://aclanthology.org/2022.lrec-1.150.pdf).
|
| 20 |
|
| 21 |
## How to Use
|
| 22 |
FairEval outputs the error count (TP, FP, etc.) and resulting scores (Precision, Recall and F1) from a reference list of
|