Spaces:
Running
Running
| title: TRAIL | |
| emoji: 🥇 | |
| colorFrom: green | |
| colorTo: indigo | |
| sdk: gradio | |
| app_file: app.py | |
| pinned: true | |
| license: mit | |
| short_description: 'TRAIL: Trace Reasoning and Agentic Issue Localization' | |
| sdk_version: 5.19.0 | |
| # Model Performance Leaderboard | |
| This is a Hugging Face Space that hosts a leaderboard for comparing model performances across various metrics of TRAIL dataset. | |
| ## Features | |
| - **Submit Model Results**: Share your model's performance metrics | |
| - **Interactive Leaderboard**: View and sort all submissions | |
| - **Integrated Backend**: Stores all submissions with timestamp and attribution | |
| - **Customizable Metrics**: Configure which metrics to display and track | |
| ## Installation | |
| ### Setting Up Your Space | |
| 1. Upload all files to your Hugging Face Space | |
| 2. Make sure to make `start.sh` executable: | |
| ```bash | |
| chmod +x start.sh | |
| ``` | |
| 3. Configure your Space to use the `start.sh` script as the entry point | |
| ### Troubleshooting Installation Issues | |
| If you encounter JSON parsing errors: | |
| 1. Check if `models.json` exists and is a valid JSON file | |
| 2. Run `python setup.py` to regenerate configuration files | |
| 3. If problems persist, delete the `models.json` file and let the setup script create a new one | |
| ## How to Use | |
| ### Viewing the Leaderboard | |
| Navigate to the "Leaderboard" tab to see all submitted models. You can: | |
| - Sort by any metric (click on the dropdown) | |
| - Change sort order (ascending/descending) | |
| - Refresh the leaderboard for the latest submissions | |
| ### Submitting a Model | |
| 1. Go to the "Submit Model" tab | |
| 2. Fill in your model name, your name, and optional description | |
| 3. Enter values for the requested metrics | |
| 4. Click "Submit Model" | |
| ## Configuration | |
| You can customize this leaderboard by modifying the `models.json` file: | |
| ```json | |
| { | |
| "title": "TRAIL Performance Leaderboard", | |
| "description": "This leaderboard tracks and compares model performance across multiple metrics. Submit your model results to see how they stack up!", | |
| "metrics": ["accuracy", "f1_score", "precision", "recall"], | |
| "main_metric": "accuracy" | |
| } | |
| ``` | |
| - `title`: The title of your leaderboard | |
| - `description`: A description that appears at the top | |
| - `metrics`: List of metrics to track | |
| - `main_metric`: Default metric for sorting | |
| ## Technical Details | |
| This leaderboard is built using: | |
| - Gradio for the UI components | |
| - A file-based database to store submissions | |
| - Pandas for data manipulation and display | |
| ## License | |
| This project is open source and available under the MIT license. |