xeon27
commited on
Commit
·
fcd47ae
1
Parent(s):
cdca101
[WIP] Add task link in description
Browse files- src/about.py +1 -1
src/about.py
CHANGED
@@ -44,7 +44,7 @@ TITLE = """<h1 align="center" id="space-title">LLM Evaluation Leaderboard</h1>""
|
|
44 |
|
45 |
# What does your leaderboard evaluate?
|
46 |
INTRODUCTION_TEXT = """
|
47 |
-
This leaderboard presents the performance of selected LLM models on a set of tasks. The tasks are divided into two categories: base and agentic. The base tasks are ARC-Easy, ARC-Challenge, DROP, WinoGrande, GSM8K, HellaSwag, HumanEval, IFEval, MATH, MMLU, MMLU-Pro, GPQA-Diamond. The agentic tasks are GAIA and GDM-InterCode-CTF.
|
48 |
"""
|
49 |
|
50 |
# Which evaluations are you running? how can people reproduce what you have?
|
|
|
44 |
|
45 |
# What does your leaderboard evaluate?
|
46 |
INTRODUCTION_TEXT = """
|
47 |
+
This leaderboard presents the performance of selected LLM models on a set of tasks. The tasks are divided into two categories: base and agentic. The base tasks are: [ARC-Easy]("https://github.com/UKGovernmentBEIS/inspect_evals/tree/main/src/inspect_evals/arc"), ARC-Challenge, DROP, WinoGrande, GSM8K, HellaSwag, HumanEval, IFEval, MATH, MMLU, MMLU-Pro, GPQA-Diamond. The agentic tasks are GAIA and GDM-InterCode-CTF.
|
48 |
"""
|
49 |
|
50 |
# Which evaluations are you running? how can people reproduce what you have?
|