Update app.py
Browse files
app.py
CHANGED
|
@@ -145,7 +145,6 @@ def evaluate_predictions(prediction_file, model_name, add_to_leaderboard):
|
|
| 145 |
initialize_leaderboard_file()
|
| 146 |
|
| 147 |
|
| 148 |
-
|
| 149 |
# Function to set default mode
|
| 150 |
css_tech_theme = """
|
| 151 |
body {
|
|
@@ -173,16 +172,15 @@ button {
|
|
| 173 |
padding: 10px 15px;
|
| 174 |
font-size: 14px;
|
| 175 |
cursor: pointer;
|
| 176 |
-
transition: background-color 0.3s ease
|
| 177 |
}
|
| 178 |
|
| 179 |
button:hover {
|
| 180 |
background-color: #6a1b9a;
|
| 181 |
-
transform: scale(1.05);
|
| 182 |
}
|
| 183 |
|
| 184 |
.input-row, .tab-content {
|
| 185 |
-
background-color: #
|
| 186 |
border-radius: 10px;
|
| 187 |
padding: 20px;
|
| 188 |
box-shadow: 0 2px 8px rgba(0, 0, 0, 0.1);
|
|
@@ -199,13 +197,14 @@ button:hover {
|
|
| 199 |
}
|
| 200 |
"""
|
| 201 |
|
|
|
|
| 202 |
with gr.Blocks(css=css_tech_theme) as demo:
|
| 203 |
gr.Markdown("""
|
| 204 |
# π Mobile-MMLU Benchmark Competition
|
| 205 |
### π Welcome to the Competition Overview
|
| 206 |

|
| 207 |
---
|
| 208 |
-
Welcome to the
|
| 209 |
---
|
| 210 |
""")
|
| 211 |
|
|
@@ -213,7 +212,7 @@ Welcome to the **Mobile-MMLU Benchmark Competition**. Here you can submit your p
|
|
| 213 |
with gr.TabItem("π Overview"):
|
| 214 |
gr.Markdown("""
|
| 215 |
## Overview
|
| 216 |
-
Welcome to the
|
| 217 |
---
|
| 218 |
### What is Mobile-MMLU?
|
| 219 |
Mobile-MMLU is a benchmark designed to test the capabilities of LLMs optimized for mobile use. Contribute to advancing mobile AI systems by competing to achieve the highest accuracy.
|
|
|
|
| 145 |
initialize_leaderboard_file()
|
| 146 |
|
| 147 |
|
|
|
|
| 148 |
# Function to set default mode
|
| 149 |
css_tech_theme = """
|
| 150 |
body {
|
|
|
|
| 172 |
padding: 10px 15px;
|
| 173 |
font-size: 14px;
|
| 174 |
cursor: pointer;
|
| 175 |
+
transition: background-color 0.3s ease;
|
| 176 |
}
|
| 177 |
|
| 178 |
button:hover {
|
| 179 |
background-color: #6a1b9a;
|
|
|
|
| 180 |
}
|
| 181 |
|
| 182 |
.input-row, .tab-content {
|
| 183 |
+
background-color: #f8f4fc;
|
| 184 |
border-radius: 10px;
|
| 185 |
padding: 20px;
|
| 186 |
box-shadow: 0 2px 8px rgba(0, 0, 0, 0.1);
|
|
|
|
| 197 |
}
|
| 198 |
"""
|
| 199 |
|
| 200 |
+
|
| 201 |
with gr.Blocks(css=css_tech_theme) as demo:
|
| 202 |
gr.Markdown("""
|
| 203 |
# π Mobile-MMLU Benchmark Competition
|
| 204 |
### π Welcome to the Competition Overview
|
| 205 |

|
| 206 |
---
|
| 207 |
+
Welcome to the Mobile-MMLU Benchmark Competition. Here you can submit your predictions, view the leaderboard, and track your performance.
|
| 208 |
---
|
| 209 |
""")
|
| 210 |
|
|
|
|
| 212 |
with gr.TabItem("π Overview"):
|
| 213 |
gr.Markdown("""
|
| 214 |
## Overview
|
| 215 |
+
Welcome to the Mobile-MMLU Benchmark Competition! Evaluate mobile-compatible Large Language Models (LLMs) on **16,186 scenario-based and factual questions** across **80 fields**.
|
| 216 |
---
|
| 217 |
### What is Mobile-MMLU?
|
| 218 |
Mobile-MMLU is a benchmark designed to test the capabilities of LLMs optimized for mobile use. Contribute to advancing mobile AI systems by competing to achieve the highest accuracy.
|