TRAIL / README.md
jitinpatronus's picture
Upload 23 files
0380c4f verified
|
raw
history blame
2.48 kB
---
title: TRAIL
emoji: 🥇
colorFrom: green
colorTo: indigo
sdk: gradio
app_file: app.py
pinned: true
license: mit
short_description: 'TRAIL: Trace Reasoning and Agentic Issue Localization'
sdk_version: 5.19.0
---
# Model Performance Leaderboard
This is a Hugging Face Space that hosts a leaderboard for comparing model performances across various metrics of TRAIL dataset.
## Features
- **Submit Model Results**: Share your model's performance metrics
- **Interactive Leaderboard**: View and sort all submissions
- **Integrated Backend**: Stores all submissions with timestamp and attribution
- **Customizable Metrics**: Configure which metrics to display and track
## Installation
### Setting Up Your Space
1. Upload all files to your Hugging Face Space
2. Make sure to make `start.sh` executable:
```bash
chmod +x start.sh
```
3. Configure your Space to use the `start.sh` script as the entry point
### Troubleshooting Installation Issues
If you encounter JSON parsing errors:
1. Check if `models.json` exists and is a valid JSON file
2. Run `python setup.py` to regenerate configuration files
3. If problems persist, delete the `models.json` file and let the setup script create a new one
## How to Use
### Viewing the Leaderboard
Navigate to the "Leaderboard" tab to see all submitted models. You can:
- Sort by any metric (click on the dropdown)
- Change sort order (ascending/descending)
- Refresh the leaderboard for the latest submissions
### Submitting a Model
1. Go to the "Submit Model" tab
2. Fill in your model name, your name, and optional description
3. Enter values for the requested metrics
4. Click "Submit Model"
## Configuration
You can customize this leaderboard by modifying the `models.json` file:
```json
{
"title": "TRAIL Performance Leaderboard",
"description": "This leaderboard tracks and compares model performance across multiple metrics. Submit your model results to see how they stack up!",
"metrics": ["accuracy", "f1_score", "precision", "recall"],
"main_metric": "accuracy"
}
```
- `title`: The title of your leaderboard
- `description`: A description that appears at the top
- `metrics`: List of metrics to track
- `main_metric`: Default metric for sorting
## Technical Details
This leaderboard is built using:
- Gradio for the UI components
- A file-based database to store submissions
- Pandas for data manipulation and display
## License
This project is open source and available under the MIT license.