Adding Error Bars to Evals: A Statistical Approach to Language Model Evaluations
Abstract
Evaluations are critical for understanding the capabilities of large language models (LLMs). Fundamentally, evaluations are experiments; but the literature on evaluations has largely ignored the literature from other sciences on experiment analysis and planning. This article shows researchers with some training in statistics how to think about and analyze data from language model evaluations. Conceptualizing evaluation questions as having been drawn from an unseen super-population, we present formulas for analyzing evaluation data, measuring differences between two models, and planning an evaluation experiment. We make a number of specific recommendations for running language model evaluations and reporting experiment results in a way that minimizes statistical noise and maximizes informativeness.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Beyond the Singular: The Essential Role of Multiple Generations in Effective Benchmark Evaluation and Analysis (2025)
- Evaluation of Large Language Models via Coupled Token Generation (2025)
- Accelerating Unbiased LLM Evaluation via Synthetic Feedback (2025)
- Language Models Prefer What They Know: Relative Confidence Estimation via Confidence Preferences (2025)
- A statistically consistent measure of Semantic Variability using Language Models (2025)
- An Empirical Analysis of Uncertainty in Large Language Model Evaluations (2025)
- Logical forms complement probability in understanding language model (and human) performance (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper