File size: 1,536 Bytes
d03b4ca
 
 
 
 
 
 
 
 
4142d5d
b17e273
 
 
 
 
 
 
 
1b0ed5c
ce58580
4142d5d
 
b17e273
 
 
 
4142d5d
b17e273
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
---
title: README
emoji: 🔥
colorFrom: yellow
colorTo: purple
sdk: static
pinned: false
---

# What is this?
This repository is a demo leaderboard template. 
You can copy the leaderboard space and the two datasets (results and requests) to your org to get started with your own leaderboard!

The space does 3 things:
- stores users submissions, and sends them to the `requests` dataset
- reads the submissions depending on their status/date of creation, and launches evaluations through the `main_backend.py` file, using the Eleuther AI Harness. Results of running evaluations are then sent to `results`
- reads the results and displays them in a leaderboard.

You can also move the backend to its own space if you need, by grabbing main_backend and putting it in its own space, with a app.py which runs it every few minutes - it is probably the best solution.

# Getting started
## Defining environment variables
To get started on your own leaderboard, you will need to edit 2 files: 
- `src/envs.py` to define your own environment variable (like the org name in which this has been copied)
- `src/about.py` with the tasks and number of few_shots you want for your tasks

## Setting up fake results to initialize the leaderboard
Once this is done, you need to edit the "fake results" file to fit the format of your tasks: in the sub dictionary `results`, replace task_name1 and metric_name by the correct values you defined in Tasks above.
```
    "results": {
        "task_name1": {
            "metric_name": 0
        }
    }
```