Spaces:
Running
Running
Commit
·
b17e273
1
Parent(s):
d03b4ca
Update README.md
Browse files
README.md
CHANGED
|
@@ -7,4 +7,23 @@ sdk: static
|
|
| 7 |
pinned: false
|
| 8 |
---
|
| 9 |
|
| 10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
pinned: false
|
| 8 |
---
|
| 9 |
|
| 10 |
+
This repository is a demo leaderboard template.
|
| 11 |
+
You can copy the leaderboard space and the two datasets (results and requests) to your org to get started with your own leaderboard!
|
| 12 |
+
|
| 13 |
+
The space does 3 things:
|
| 14 |
+
- stores users submissions, and sends them to the `requests` dataset
|
| 15 |
+
- reads the submissions depending on their status/date of creation, and launches evaluations through the `main_backend.py` file, using the Eleuther AI Harness. Results of running evaluations are then sent to `results`
|
| 16 |
+
- reads the results and displays them in a leaderboard.
|
| 17 |
+
|
| 18 |
+
To get started on your own leaderboard, you will need to edit 2 files:
|
| 19 |
+
- `src/envs.py` to define your own environment variable (like the org name in which this has been copied)
|
| 20 |
+
- `src/about.py` with the tasks and number of few_shots you want for your tasks
|
| 21 |
+
|
| 22 |
+
Once this is done, you need to edit the "fake results" file to fit the format of your tasks: in the sub dictionary `results`, replace task_name1 and metric_name by the correct values you defined in Tasks above.
|
| 23 |
+
```
|
| 24 |
+
"results": {
|
| 25 |
+
"task_name1": {
|
| 26 |
+
"metric_name": 0
|
| 27 |
+
}
|
| 28 |
+
}
|
| 29 |
+
```
|