Spaces:
Running
Running
Commit
·
6616540
1
Parent(s):
9c195eb
[FIX] Model size filter
Browse files- app.py +2 -1
- src/display/utils.py +1 -1
- src/leaderboard/read_evals.py +1 -0
app.py
CHANGED
|
@@ -167,6 +167,7 @@ def update_table(
|
|
| 167 |
precision_query: str = None,
|
| 168 |
show_deleted: bool = False,
|
| 169 |
):
|
|
|
|
| 170 |
filtered_df = filter_models(hidden_df, type_query, domain_specific_query, size_query, precision_query, show_deleted)
|
| 171 |
# breakpoint()
|
| 172 |
filtered_df = filter_queries(query, filtered_df)
|
|
@@ -245,7 +246,7 @@ def filter_models(
|
|
| 245 |
if AutoEvalColumn.precision.name in df.columns:
|
| 246 |
filtered_df = filtered_df.loc[df[AutoEvalColumn.precision.name].isin(precision_query + ["None"])]
|
| 247 |
|
| 248 |
-
if size_query is not None
|
| 249 |
numeric_interval = pd.IntervalIndex(sorted([NUMERIC_INTERVALS[s] for s in size_query]))
|
| 250 |
params_column = pd.to_numeric(df[AutoEvalColumn.params.name], errors="coerce")
|
| 251 |
mask = params_column.apply(lambda x: any(numeric_interval.contains(x)))
|
|
|
|
| 167 |
precision_query: str = None,
|
| 168 |
show_deleted: bool = False,
|
| 169 |
):
|
| 170 |
+
# breakpoint()
|
| 171 |
filtered_df = filter_models(hidden_df, type_query, domain_specific_query, size_query, precision_query, show_deleted)
|
| 172 |
# breakpoint()
|
| 173 |
filtered_df = filter_queries(query, filtered_df)
|
|
|
|
| 246 |
if AutoEvalColumn.precision.name in df.columns:
|
| 247 |
filtered_df = filtered_df.loc[df[AutoEvalColumn.precision.name].isin(precision_query + ["None"])]
|
| 248 |
|
| 249 |
+
if size_query is not None:
|
| 250 |
numeric_interval = pd.IntervalIndex(sorted([NUMERIC_INTERVALS[s] for s in size_query]))
|
| 251 |
params_column = pd.to_numeric(df[AutoEvalColumn.params.name], errors="coerce")
|
| 252 |
mask = params_column.apply(lambda x: any(numeric_interval.contains(x)))
|
src/display/utils.py
CHANGED
|
@@ -248,7 +248,7 @@ CLOSED_ENDED_ARABIC_BENCHMARK_COLS = [t.value.col_name for t in ClosedEndedArabi
|
|
| 248 |
# CROSS_EXAMINATION_BENCHMARK_COLS = [t.value.col_name for t in CrossExaminationTasks]
|
| 249 |
|
| 250 |
NUMERIC_INTERVALS = {
|
| 251 |
-
"?": pd.Interval(-
|
| 252 |
"~1.5": pd.Interval(0, 2, closed="right"),
|
| 253 |
"~3": pd.Interval(2, 4, closed="right"),
|
| 254 |
"~7": pd.Interval(4, 9, closed="right"),
|
|
|
|
| 248 |
# CROSS_EXAMINATION_BENCHMARK_COLS = [t.value.col_name for t in CrossExaminationTasks]
|
| 249 |
|
| 250 |
NUMERIC_INTERVALS = {
|
| 251 |
+
"?": pd.Interval(-100, 0, closed="right"),
|
| 252 |
"~1.5": pd.Interval(0, 2, closed="right"),
|
| 253 |
"~3": pd.Interval(2, 4, closed="right"),
|
| 254 |
"~7": pd.Interval(4, 9, closed="right"),
|
src/leaderboard/read_evals.py
CHANGED
|
@@ -63,6 +63,7 @@ class EvalResult:
|
|
| 63 |
model_type = ModelType.from_str(config.get("model_type", ""))
|
| 64 |
license = config.get("license", "?")
|
| 65 |
num_params = config.get("num_params", "?")
|
|
|
|
| 66 |
display_result = config.get("display_result", True)
|
| 67 |
display_result = False if display_result=="False" else True
|
| 68 |
|
|
|
|
| 63 |
model_type = ModelType.from_str(config.get("model_type", ""))
|
| 64 |
license = config.get("license", "?")
|
| 65 |
num_params = config.get("num_params", "?")
|
| 66 |
+
num_params = -1 if num_params == "?" or num_params is None or isinstance(num_params, float) and math.isnan(num_params) else num_params
|
| 67 |
display_result = config.get("display_result", True)
|
| 68 |
display_result = False if display_result=="False" else True
|
| 69 |
|