Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
eduagarcia
commited on
Commit
•
b234783
1
Parent(s):
4445ad2
Add hidden option
Browse files
src/leaderboard/read_evals.py
CHANGED
@@ -38,6 +38,7 @@ class EvalResult:
|
|
38 |
json_filename: str = None
|
39 |
eval_time: float = 0.0
|
40 |
original_benchmark_average: float = None
|
|
|
41 |
|
42 |
@classmethod
|
43 |
def init_from_json_file(self, json_filepath, is_original=False):
|
@@ -125,6 +126,7 @@ class EvalResult:
|
|
125 |
self.date = request.get("submitted_time", "")
|
126 |
self.architecture = request.get("architectures", "Unknown")
|
127 |
self.status = request.get("status", "FAILED")
|
|
|
128 |
except Exception as e:
|
129 |
self.status = "FAILED"
|
130 |
print(f"Could not find request file for {self.org}/{self.model}")
|
@@ -163,7 +165,7 @@ class EvalResult:
|
|
163 |
AutoEvalColumn.merged.name: "merge" in self.tags if self.tags else False,
|
164 |
AutoEvalColumn.moe.name: ("moe" in self.tags if self.tags else False) or "moe" in self.full_model.lower(),
|
165 |
AutoEvalColumn.flagged.name: self.flagged,
|
166 |
-
AutoEvalColumn.eval_time.name: self.eval_time
|
167 |
}
|
168 |
|
169 |
for task in Tasks:
|
@@ -236,7 +238,7 @@ def get_raw_eval_results(results_path: str, requests_path: str, dynamic_path: st
|
|
236 |
results = []
|
237 |
for v in eval_results.values():
|
238 |
try:
|
239 |
-
if v.status == "FINISHED":
|
240 |
v.to_dict() # we test if the dict version is complete
|
241 |
results.append(v)
|
242 |
except KeyError: # not all eval values present
|
|
|
38 |
json_filename: str = None
|
39 |
eval_time: float = 0.0
|
40 |
original_benchmark_average: float = None
|
41 |
+
hidden: bool = False # Do not show on the leaderboard
|
42 |
|
43 |
@classmethod
|
44 |
def init_from_json_file(self, json_filepath, is_original=False):
|
|
|
126 |
self.date = request.get("submitted_time", "")
|
127 |
self.architecture = request.get("architectures", "Unknown")
|
128 |
self.status = request.get("status", "FAILED")
|
129 |
+
self.hidden = request.get("hidden", False)
|
130 |
except Exception as e:
|
131 |
self.status = "FAILED"
|
132 |
print(f"Could not find request file for {self.org}/{self.model}")
|
|
|
165 |
AutoEvalColumn.merged.name: "merge" in self.tags if self.tags else False,
|
166 |
AutoEvalColumn.moe.name: ("moe" in self.tags if self.tags else False) or "moe" in self.full_model.lower(),
|
167 |
AutoEvalColumn.flagged.name: self.flagged,
|
168 |
+
AutoEvalColumn.eval_time.name: self.eval_time
|
169 |
}
|
170 |
|
171 |
for task in Tasks:
|
|
|
238 |
results = []
|
239 |
for v in eval_results.values():
|
240 |
try:
|
241 |
+
if v.status == "FINISHED" and not v.hidden:
|
242 |
v.to_dict() # we test if the dict version is complete
|
243 |
results.append(v)
|
244 |
except KeyError: # not all eval values present
|