|
import json |
|
from collections import defaultdict |
|
from pathlib import Path |
|
|
|
import numpy as np |
|
import pandas as pd |
|
import gradio as gr |
|
from pandas import DataFrame |
|
from pandas.io.formats.style import Styler |
|
|
|
from content import * |
|
|
|
ARC = "arc" |
|
HELLASWAG = "hellaswag" |
|
MMLU = "mmlu" |
|
TRUTHFULQA = "truthfulqa" |
|
BENCHMARKS = [ARC, HELLASWAG, MMLU, TRUTHFULQA] |
|
|
|
METRICS = ["acc_norm", "acc_norm", "acc_norm", "mc2"] |
|
|
|
|
|
def collect_results() -> dict[tuple[str, str], dict[str, float]]: |
|
""" |
|
Collects results from the evals folder and returns a dictionary of results |
|
:return: a dictionary of results where the keys are typles of (model_name, language) and the values are |
|
dictionaries of the form {benchmark_name: performance_score} |
|
""" |
|
performance_dict = defaultdict(dict) |
|
for pfin in Path("evals").rglob("*.json"): |
|
data = json.loads(pfin.read_text(encoding="utf-8")) |
|
if "results" not in data or "config" not in data: |
|
continue |
|
results = data["results"] |
|
config = data["config"] |
|
if "model_args" not in config: |
|
continue |
|
|
|
model_args = config["model_args"].split(",") |
|
pretrained = [x for x in model_args if x.startswith("pretrained=")] |
|
if len(pretrained) != 1: |
|
continue |
|
pretrained = pretrained[0].split("=")[1] |
|
pretrained = pretrained.split("/")[-1] |
|
|
|
for lang_task, perfs in results.items(): |
|
task, lang = lang_task.split("_") |
|
assert task in BENCHMARKS |
|
|
|
if lang and task: |
|
metric = METRICS[BENCHMARKS.index(task)] |
|
p = round(perfs[metric] * 100, 1) |
|
performance_dict[(pretrained, lang)][task] = p |
|
|
|
return dict(performance_dict) |
|
|
|
|
|
def build_performance_df(performance_dict: dict[tuple[str, str], dict[str, float]]) -> DataFrame: |
|
""" |
|
Builds a dataframe from the performance dictionary |
|
:param performance_dict: a dictionary of results where the keys are typles of (model_name, language) and the values are |
|
dictionaries of the form {benchmark_name: performance_score} |
|
:return: a pd.DataFrame that has as rows the model names and as columns the benchmarks |
|
""" |
|
data = [] |
|
for (pretrained, lang), perfs in performance_dict.items(): |
|
arc_perf = perfs.get(ARC, 0.0) |
|
hellaswag_perf = perfs.get(HELLASWAG, 0.0) |
|
mmlu_perf = perfs.get(MMLU, 0.0) |
|
truthfulqa_perf = perfs.get(TRUTHFULQA, 0.0) |
|
|
|
avg = round((arc_perf + hellaswag_perf + mmlu_perf + truthfulqa_perf) / 4, 1) |
|
row = [pretrained, avg, arc_perf, hellaswag_perf, mmlu_perf, truthfulqa_perf] |
|
data.append(row) |
|
|
|
df = pd.DataFrame.from_records(data, columns=COLS) |
|
df = df.sort_values(by=[AVERAGE_COL], ascending=False) |
|
return df |
|
|
|
|
|
def style_df(df: DataFrame) -> Styler: |
|
""" |
|
Styles the dataframe by rounding to two decimals and putting the max value in bold per column |
|
:param df: the dataframe to style |
|
:return: the Styler |
|
""" |
|
styler = df.style.format("{:.2f}", subset=df.columns[1:]) |
|
|
|
def highlight_max(col): |
|
return np.where(col == np.nanmax(col.to_numpy()), "font-weight: bold;", None) |
|
|
|
styler = styler.apply(highlight_max, axis=1, subset=df.columns[1:]) |
|
styler = styler.hide() |
|
return styler |
|
|
|
|
|
MODEL_COL = "Model" |
|
AVERAGE_COL = "Average" |
|
ARC_COL = "ARC (25-shot)" |
|
HELLASWAG_COL = "HellaSwag (10-shot)️" |
|
MMLU_COL = "MMLU (5-shot)" |
|
TRUTHFULQA_COL = "TruthfulQA (0-shot)" |
|
|
|
COLS = [MODEL_COL, AVERAGE_COL, ARC_COL, HELLASWAG_COL, MMLU_COL, TRUTHFULQA_COL] |
|
TYPES = ["str", "number", "number", "number", "number", "number"] |
|
|
|
results = collect_results() |
|
original_df = build_performance_df(results) |
|
styled_df = style_df(original_df) |
|
with gr.Blocks() as demo: |
|
gr.HTML(TITLE) |
|
gr.Markdown(INTRO_TEXT) |
|
|
|
gr.Markdown("## Leaderboard\nOnly representative for the Dutch version (`*_nl`) of the benchmarks!") |
|
gr.components.Dataframe( |
|
value=original_df, |
|
headers=COLS, |
|
datatype=TYPES, |
|
elem_id="leaderboard-table", |
|
) |
|
|
|
gr.Markdown("## LaTeX") |
|
gr.Code(styled_df.to_latex(convert_css=True)) |
|
|
|
gr.Markdown(CREDIT, elem_classes="markdown-text") |
|
gr.Markdown(CITATION, elem_classes="markdown-text") |
|
|
|
if __name__ == '__main__': |
|
demo.launch() |
|
|
|
|