Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
·
b234783
1
Parent(s):
4445ad2
Add hidden option
Browse files
src/leaderboard/read_evals.py
CHANGED
|
@@ -38,6 +38,7 @@ class EvalResult:
|
|
| 38 |
json_filename: str = None
|
| 39 |
eval_time: float = 0.0
|
| 40 |
original_benchmark_average: float = None
|
|
|
|
| 41 |
|
| 42 |
@classmethod
|
| 43 |
def init_from_json_file(self, json_filepath, is_original=False):
|
|
@@ -125,6 +126,7 @@ class EvalResult:
|
|
| 125 |
self.date = request.get("submitted_time", "")
|
| 126 |
self.architecture = request.get("architectures", "Unknown")
|
| 127 |
self.status = request.get("status", "FAILED")
|
|
|
|
| 128 |
except Exception as e:
|
| 129 |
self.status = "FAILED"
|
| 130 |
print(f"Could not find request file for {self.org}/{self.model}")
|
|
@@ -163,7 +165,7 @@ class EvalResult:
|
|
| 163 |
AutoEvalColumn.merged.name: "merge" in self.tags if self.tags else False,
|
| 164 |
AutoEvalColumn.moe.name: ("moe" in self.tags if self.tags else False) or "moe" in self.full_model.lower(),
|
| 165 |
AutoEvalColumn.flagged.name: self.flagged,
|
| 166 |
-
AutoEvalColumn.eval_time.name: self.eval_time
|
| 167 |
}
|
| 168 |
|
| 169 |
for task in Tasks:
|
|
@@ -236,7 +238,7 @@ def get_raw_eval_results(results_path: str, requests_path: str, dynamic_path: st
|
|
| 236 |
results = []
|
| 237 |
for v in eval_results.values():
|
| 238 |
try:
|
| 239 |
-
if v.status == "FINISHED":
|
| 240 |
v.to_dict() # we test if the dict version is complete
|
| 241 |
results.append(v)
|
| 242 |
except KeyError: # not all eval values present
|
|
|
|
| 38 |
json_filename: str = None
|
| 39 |
eval_time: float = 0.0
|
| 40 |
original_benchmark_average: float = None
|
| 41 |
+
hidden: bool = False # Do not show on the leaderboard
|
| 42 |
|
| 43 |
@classmethod
|
| 44 |
def init_from_json_file(self, json_filepath, is_original=False):
|
|
|
|
| 126 |
self.date = request.get("submitted_time", "")
|
| 127 |
self.architecture = request.get("architectures", "Unknown")
|
| 128 |
self.status = request.get("status", "FAILED")
|
| 129 |
+
self.hidden = request.get("hidden", False)
|
| 130 |
except Exception as e:
|
| 131 |
self.status = "FAILED"
|
| 132 |
print(f"Could not find request file for {self.org}/{self.model}")
|
|
|
|
| 165 |
AutoEvalColumn.merged.name: "merge" in self.tags if self.tags else False,
|
| 166 |
AutoEvalColumn.moe.name: ("moe" in self.tags if self.tags else False) or "moe" in self.full_model.lower(),
|
| 167 |
AutoEvalColumn.flagged.name: self.flagged,
|
| 168 |
+
AutoEvalColumn.eval_time.name: self.eval_time
|
| 169 |
}
|
| 170 |
|
| 171 |
for task in Tasks:
|
|
|
|
| 238 |
results = []
|
| 239 |
for v in eval_results.values():
|
| 240 |
try:
|
| 241 |
+
if v.status == "FINISHED" and not v.hidden:
|
| 242 |
v.to_dict() # we test if the dict version is complete
|
| 243 |
results.append(v)
|
| 244 |
except KeyError: # not all eval values present
|