llm-perf-leaderboard / src /flashattentionv2.py
IlyasMoutawwakil's picture
update
0232cf1
raw
history blame
4.32 kB
import gradio as gr
import pandas as pd
import plotly.express as px
FLASHATTENTIONV2_DATA = [
# open llm
"Model πŸ€—",
"DType πŸ“₯",
"Backend 🏭",
"Params (B)",
"Architecture πŸ›οΈ",
"Open LLM Score (%)",
# deployment settings
"DType πŸ“₯",
"Backend 🏭",
"Optimization πŸ› οΈ",
"Quantization πŸ—œοΈ",
"Optimization πŸ› οΈ FlashAttentionV2",
# primary measurements
"Prefill (s)",
"Prefill (s) FlashAttentionV2",
"Decode (tokens/s)",
"Decode (tokens/s) FlashAttentionV2",
"End-to-End (tokens/s)",
"End-to-End (tokens/s) FlashAttentionV2",
# speedups
"Prefill Speedup (%)",
"Decode Speedup (%)",
]
def get_fa2_df(llm_perf_df):
copy_df = llm_perf_df.copy()
# seperate original model experiments from FlashAttentionV2 experiments
original_df = copy_df[(copy_df["Optimization πŸ› οΈ"] == "None") & (copy_df["DType πŸ“₯"] == "float16")]
fa2_df = copy_df[(copy_df["Optimization πŸ› οΈ"] == "FlashAttentionV2") & (copy_df["DType πŸ“₯"] == "float16")]
# merge the two dataframes
fa2_df = pd.merge(
original_df,
fa2_df,
on=["Model πŸ€—", "Quantization πŸ—œοΈ"],
suffixes=["", " FlashAttentionV2"],
)
# compute speedups
fa2_df["Prefill Speedup (%)"] = ((fa2_df["Prefill (s)"] / fa2_df["Prefill (s) FlashAttentionV2"]) * 100).round(
2
) - 100
fa2_df["Decode Speedup (%)"] = (
(fa2_df["Decode (tokens/s) FlashAttentionV2"] / fa2_df["Decode (tokens/s)"]) * 100
).round(2) - 100
# filter speedups > 1000%
fa2_df = fa2_df[fa2_df["Prefill Speedup (%)"] < 1000]
fa2_df = fa2_df[fa2_df["Decode Speedup (%)"] < 1000]
return fa2_df
def get_fa2_decode_fig(llm_perf_df):
fa2_df = get_fa2_df(llm_perf_df)
# plot
decode_fig = px.box(
fa2_df,
x="Architecture πŸ›οΈ",
y="Decode Speedup (%)",
color_discrete_sequence=px.colors.qualitative.Light24,
custom_data=FLASHATTENTIONV2_DATA,
color="Quantization πŸ—œοΈ",
points="all",
)
# add hover data
decode_fig.update_traces(
hovertemplate="<br>".join(
[f"<b>{column}:</b> %{{customdata[{i}]}}" for i, column in enumerate(FLASHATTENTIONV2_DATA)]
)
)
# add layout
decode_fig.update_layout(
title={
"text": "Decode Speedup per Architecture, Compared To Non-Optimized Model",
"y": 0.95,
"x": 0.5,
"xanchor": "center",
"yanchor": "top",
},
xaxis_title="LLM Architecture",
yaxis_title="Decode Speedup (%)",
legend_title="Quantization Scheme",
width=1200,
height=600,
)
return decode_fig
def get_fa2_prefill_fig(llm_perf_df):
fa2_df = get_fa2_df(llm_perf_df)
# plot
prefill_fig = px.box(
fa2_df,
x="Architecture πŸ›οΈ",
y="Prefill Speedup (%)",
color_discrete_sequence=px.colors.qualitative.Light24,
custom_data=FLASHATTENTIONV2_DATA,
color="Quantization πŸ—œοΈ",
points="all",
)
# add hover data
prefill_fig.update_traces(
hovertemplate="<br>".join(
[f"<b>{column}:</b> %{{customdata[{i}]}}" for i, column in enumerate(FLASHATTENTIONV2_DATA)]
)
)
# add layout
prefill_fig.update_layout(
title={
"text": "Prefill Speedup per Architecture, Compared To Non-Optimized Model",
"y": 0.95,
"x": 0.5,
"xanchor": "center",
"yanchor": "top",
},
xaxis_title="LLM Architecture",
yaxis_title="Prefill Speedup (%)",
legend_title="Quantization Scheme",
width=1200,
height=600,
)
return prefill_fig
def create_fa2_plots(llm_perf_df):
# descriptive text
gr.HTML("πŸ‘† Hover over the points πŸ‘† for additional information.", elem_id="text")
# get figures
prefill_fig = get_fa2_prefill_fig(llm_perf_df)
decode_fig = get_fa2_decode_fig(llm_perf_df)
# create plots
prefill_plot = gr.components.Plot(value=prefill_fig, elem_id="plot", show_label=False)
decode_plot = gr.components.Plot(value=decode_fig, elem_id="plot", show_label=False)
return prefill_plot, decode_plot