TRAIL / README.md
jitinpatronus's picture
Upload 23 files
0380c4f verified
|
raw
history blame
2.48 kB
metadata
title: TRAIL
emoji: 🥇
colorFrom: green
colorTo: indigo
sdk: gradio
app_file: app.py
pinned: true
license: mit
short_description: 'TRAIL: Trace Reasoning and Agentic Issue Localization'
sdk_version: 5.19.0

Model Performance Leaderboard

This is a Hugging Face Space that hosts a leaderboard for comparing model performances across various metrics of TRAIL dataset.

Features

  • Submit Model Results: Share your model's performance metrics
  • Interactive Leaderboard: View and sort all submissions
  • Integrated Backend: Stores all submissions with timestamp and attribution
  • Customizable Metrics: Configure which metrics to display and track

Installation

Setting Up Your Space

  1. Upload all files to your Hugging Face Space
  2. Make sure to make start.sh executable:
    chmod +x start.sh
    
  3. Configure your Space to use the start.sh script as the entry point

Troubleshooting Installation Issues

If you encounter JSON parsing errors:

  1. Check if models.json exists and is a valid JSON file
  2. Run python setup.py to regenerate configuration files
  3. If problems persist, delete the models.json file and let the setup script create a new one

How to Use

Viewing the Leaderboard

Navigate to the "Leaderboard" tab to see all submitted models. You can:

  • Sort by any metric (click on the dropdown)
  • Change sort order (ascending/descending)
  • Refresh the leaderboard for the latest submissions

Submitting a Model

  1. Go to the "Submit Model" tab
  2. Fill in your model name, your name, and optional description
  3. Enter values for the requested metrics
  4. Click "Submit Model"

Configuration

You can customize this leaderboard by modifying the models.json file:

{
  "title": "TRAIL Performance Leaderboard",
  "description": "This leaderboard tracks and compares model performance across multiple metrics. Submit your model results to see how they stack up!",
  "metrics": ["accuracy", "f1_score", "precision", "recall"],
  "main_metric": "accuracy"
}
  • title: The title of your leaderboard
  • description: A description that appears at the top
  • metrics: List of metrics to track
  • main_metric: Default metric for sorting

Technical Details

This leaderboard is built using:

  • Gradio for the UI components
  • A file-based database to store submissions
  • Pandas for data manipulation and display

License

This project is open source and available under the MIT license.