Spaces:
Running
on
Zero
Running
on
Zero
DongfuJiang
commited on
Commit
•
0b4b1e4
1
Parent(s):
a4a785c
update model registry
Browse files- arena_elo/elo_rating/clean_battle_data.py +3 -1
- arena_elo/elo_rating/elo_analysis.py +1 -1
- arena_elo/elo_rating/generate_leaderboard.py +14 -14
- arena_elo/elo_rating/model_registry.py +0 -578
- arena_elo/elo_rating/utils.py +12 -4
- arena_elo/results/20240809/elo_results_image_editing.pkl +2 -2
- arena_elo/results/20240809/image_editing_leaderboard.csv +6 -6
- arena_elo/results/20240819/clean_battle_video_generation.json +0 -0
- arena_elo/results/20240819/elo_results_video_generation.pkl +2 -2
- arena_elo/results/20240819/video_generation_leaderboard.csv +5 -5
- arena_elo/results/20240820/clean_battle_t2i_generation.json +0 -0
- arena_elo/results/20240820/elo_results_t2i_generation.pkl +2 -2
- arena_elo/results/20240820/t2i_generation_leaderboard.csv +17 -17
- arena_elo/results/latest/clean_battle_t2i_generation.json +0 -0
- arena_elo/results/latest/clean_battle_video_generation.json +0 -0
- arena_elo/results/latest/elo_results_image_editing.pkl +2 -2
- arena_elo/results/latest/elo_results_t2i_generation.pkl +2 -2
- arena_elo/results/latest/elo_results_video_generation.pkl +2 -2
- arena_elo/results/latest/image_editing_leaderboard.csv +6 -6
- arena_elo/results/latest/t2i_generation_leaderboard.csv +17 -17
- arena_elo/results/latest/video_generation_leaderboard.csv +4 -4
- arena_elo/update_elo_rating.sh +0 -3
- model/model_registry.py +159 -10
- model/models/__init__.py +7 -0
arena_elo/elo_rating/clean_battle_data.py
CHANGED
@@ -18,7 +18,7 @@ ImageFile.LOAD_TRUNCATED_IMAGES = True
|
|
18 |
from tqdm import tqdm
|
19 |
|
20 |
from .basic_stats import get_log_files, NUM_SERVERS, LOG_ROOT_DIR
|
21 |
-
from .utils import detect_language, get_time_stamp_from_date
|
22 |
|
23 |
VOTES = ["tievote", "leftvote", "rightvote", "bothbad_vote"]
|
24 |
|
@@ -54,6 +54,8 @@ def replace_model_name(old_name, tstamp):
|
|
54 |
if "Flux" in old_name:
|
55 |
print(f"Invalid model names: {old_name}")
|
56 |
exit(1)
|
|
|
|
|
57 |
return old_name
|
58 |
|
59 |
|
|
|
18 |
from tqdm import tqdm
|
19 |
|
20 |
from .basic_stats import get_log_files, NUM_SERVERS, LOG_ROOT_DIR
|
21 |
+
from .utils import detect_language, get_time_stamp_from_date, get_model_info
|
22 |
|
23 |
VOTES = ["tievote", "leftvote", "rightvote", "bothbad_vote"]
|
24 |
|
|
|
54 |
if "Flux" in old_name:
|
55 |
print(f"Invalid model names: {old_name}")
|
56 |
exit(1)
|
57 |
+
model_info = get_model_info(old_name)
|
58 |
+
old_name = model_info.simple_name
|
59 |
return old_name
|
60 |
|
61 |
|
arena_elo/elo_rating/elo_analysis.py
CHANGED
@@ -11,9 +11,9 @@ import pandas as pd
|
|
11 |
import plotly.express as px
|
12 |
from tqdm import tqdm
|
13 |
|
14 |
-
from .model_registry import get_model_info
|
15 |
from .basic_stats import get_log_files
|
16 |
from .clean_battle_data import clean_battle_data
|
|
|
17 |
|
18 |
pd.options.display.float_format = "{:.2f}".format
|
19 |
|
|
|
11 |
import plotly.express as px
|
12 |
from tqdm import tqdm
|
13 |
|
|
|
14 |
from .basic_stats import get_log_files
|
15 |
from .clean_battle_data import clean_battle_data
|
16 |
+
from .utils import get_model_info
|
17 |
|
18 |
pd.options.display.float_format = "{:.2f}".format
|
19 |
|
arena_elo/elo_rating/generate_leaderboard.py
CHANGED
@@ -2,15 +2,12 @@ import fire
|
|
2 |
import json
|
3 |
import pandas as pd
|
4 |
import pickle
|
5 |
-
|
6 |
|
7 |
def main(
|
8 |
-
|
9 |
-
|
10 |
-
output_csv: str
|
11 |
):
|
12 |
-
model_info = json.load(open(model_info_file))
|
13 |
-
|
14 |
with open(elo_rating_pkl, "rb") as fin:
|
15 |
elo_rating_results = pickle.load(fin)
|
16 |
|
@@ -19,20 +16,23 @@ def main(
|
|
19 |
anony_leaderboard_data = anony_elo_rating_results["leaderboard_table_df"]
|
20 |
full_leaderboard_data = full_elo_rating_results["leaderboard_table_df"]
|
21 |
|
|
|
22 |
# Model,MT-bench (score),Arena Elo rating,MMLU,License,Link
|
23 |
fields = ["key", "Model", "Arena Elo rating (anony)", "Arena Elo rating (full)", "License", "Organization", "Link"]
|
24 |
# set Organization and license to empty for now
|
25 |
all_models = anony_leaderboard_data.index.tolist()
|
26 |
|
|
|
27 |
for model in all_models:
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
|
|
36 |
|
37 |
if model in anony_leaderboard_data.index:
|
38 |
model_info[model]["Arena Elo rating (anony)"] = anony_leaderboard_data.loc[model, "rating"]
|
|
|
2 |
import json
|
3 |
import pandas as pd
|
4 |
import pickle
|
5 |
+
from .utils import get_model_info
|
6 |
|
7 |
def main(
|
8 |
+
elo_rating_pkl: str,
|
9 |
+
output_csv: str
|
|
|
10 |
):
|
|
|
|
|
11 |
with open(elo_rating_pkl, "rb") as fin:
|
12 |
elo_rating_results = pickle.load(fin)
|
13 |
|
|
|
16 |
anony_leaderboard_data = anony_elo_rating_results["leaderboard_table_df"]
|
17 |
full_leaderboard_data = full_elo_rating_results["leaderboard_table_df"]
|
18 |
|
19 |
+
print(anony_leaderboard_data)
|
20 |
# Model,MT-bench (score),Arena Elo rating,MMLU,License,Link
|
21 |
fields = ["key", "Model", "Arena Elo rating (anony)", "Arena Elo rating (full)", "License", "Organization", "Link"]
|
22 |
# set Organization and license to empty for now
|
23 |
all_models = anony_leaderboard_data.index.tolist()
|
24 |
|
25 |
+
model_info = {}
|
26 |
for model in all_models:
|
27 |
+
|
28 |
+
registered_model_info = get_model_info(model)
|
29 |
+
model_info[model] = {
|
30 |
+
"key": model,
|
31 |
+
"Model": model,
|
32 |
+
"License": registered_model_info.license,
|
33 |
+
"Organization": registered_model_info.organization,
|
34 |
+
"Link": registered_model_info.link
|
35 |
+
}
|
36 |
|
37 |
if model in anony_leaderboard_data.index:
|
38 |
model_info[model]["Arena Elo rating (anony)"] = anony_leaderboard_data.loc[model, "rating"]
|
arena_elo/elo_rating/model_registry.py
DELETED
@@ -1,578 +0,0 @@
|
|
1 |
-
"""Additional information of the models."""
|
2 |
-
from collections import namedtuple, OrderedDict
|
3 |
-
from typing import List
|
4 |
-
|
5 |
-
|
6 |
-
ModelInfo = namedtuple("ModelInfo", ["simple_name", "link", "description"])
|
7 |
-
|
8 |
-
|
9 |
-
model_info = OrderedDict()
|
10 |
-
|
11 |
-
|
12 |
-
def register_model_info(
|
13 |
-
full_names: List[str], simple_name: str, link: str, description: str
|
14 |
-
):
|
15 |
-
info = ModelInfo(simple_name, link, description)
|
16 |
-
|
17 |
-
for full_name in full_names:
|
18 |
-
model_info[full_name] = info
|
19 |
-
|
20 |
-
|
21 |
-
def get_model_info(name: str) -> ModelInfo:
|
22 |
-
if name in model_info:
|
23 |
-
return model_info[name]
|
24 |
-
else:
|
25 |
-
# To fix this, please use `register_model_info` to register your model
|
26 |
-
return ModelInfo(
|
27 |
-
name, "", "Register the description at arena.model/model_registry.py"
|
28 |
-
)
|
29 |
-
|
30 |
-
|
31 |
-
register_model_info(
|
32 |
-
[
|
33 |
-
"IEITYuan/Yuan2-2B-Janus-hf",
|
34 |
-
"IEITYuan/Yuan2-2B-hf",
|
35 |
-
"IEITYuan/Yuan2-51B-hf",
|
36 |
-
"IEITYuan/Yuan2-102B-hf",
|
37 |
-
],
|
38 |
-
"IEIT-Yuan2",
|
39 |
-
"https://github.com/IEIT-Yuan/Yuan-2.0",
|
40 |
-
"Yuan2.0 is a new generation Fundamental Large Language Model developed by IEIT System.",
|
41 |
-
)
|
42 |
-
|
43 |
-
register_model_info(
|
44 |
-
["mixtral-8x7b-instruct-v0.1", "mistral-7b-instruct"],
|
45 |
-
"Mixtral of experts",
|
46 |
-
"https://mistral.ai/news/mixtral-of-experts/",
|
47 |
-
"A Mixture-of-Experts model by Mistral AI",
|
48 |
-
)
|
49 |
-
|
50 |
-
register_model_info(
|
51 |
-
["gemini-pro"],
|
52 |
-
"Gemini",
|
53 |
-
"https://blog.google/technology/ai/google-gemini-pro-imagen-duet-ai-update/",
|
54 |
-
"Gemini by Google",
|
55 |
-
)
|
56 |
-
|
57 |
-
register_model_info(
|
58 |
-
["gemini-pro-vision"],
|
59 |
-
"Gemini",
|
60 |
-
"https://blog.google/technology/ai/google-gemini-pro-imagen-duet-ai-update/",
|
61 |
-
"Gemini by Google",
|
62 |
-
)
|
63 |
-
|
64 |
-
register_model_info(
|
65 |
-
["solar-10.7b-instruct-v1.0"],
|
66 |
-
"SOLAR-10.7B-Instruct",
|
67 |
-
"https://huggingface.co/upstage/SOLAR-10.7B-Instruct-v1.0",
|
68 |
-
"A model trained using depth up-scaling by Upstage AI",
|
69 |
-
)
|
70 |
-
|
71 |
-
register_model_info(
|
72 |
-
["gpt-4-turbo"],
|
73 |
-
"GPT-4-Turbo",
|
74 |
-
"https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo",
|
75 |
-
"GPT-4-Turbo by OpenAI",
|
76 |
-
)
|
77 |
-
|
78 |
-
register_model_info(
|
79 |
-
["gpt-4-vision-preview"],
|
80 |
-
"gpt-4-vision-preview",
|
81 |
-
"https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo",
|
82 |
-
"GPT-4(V) by OpenAI",
|
83 |
-
)
|
84 |
-
|
85 |
-
register_model_info(
|
86 |
-
["gpt-3.5-turbo", "gpt-3.5-turbo-0314", "gpt-3.5-turbo-0613", "gpt-3.5-turbo-1106"],
|
87 |
-
"GPT-3.5",
|
88 |
-
"https://platform.openai.com/docs/models/gpt-3-5",
|
89 |
-
"GPT-3.5-Turbo by OpenAI",
|
90 |
-
)
|
91 |
-
|
92 |
-
register_model_info(
|
93 |
-
["gpt-4", "gpt-4-0314", "gpt-4-0613"],
|
94 |
-
"GPT-4",
|
95 |
-
"https://openai.com/research/gpt-4",
|
96 |
-
"GPT-4 by OpenAI",
|
97 |
-
)
|
98 |
-
|
99 |
-
register_model_info(
|
100 |
-
["claude-2.1", "claude-2.0"],
|
101 |
-
"Claude",
|
102 |
-
"https://www.anthropic.com/index/claude-2",
|
103 |
-
"Claude 2 by Anthropic",
|
104 |
-
)
|
105 |
-
|
106 |
-
register_model_info(
|
107 |
-
["claude-1"],
|
108 |
-
"Claude",
|
109 |
-
"https://www.anthropic.com/index/introducing-claude",
|
110 |
-
"Claude 1 by Anthropic",
|
111 |
-
)
|
112 |
-
|
113 |
-
register_model_info(
|
114 |
-
["claude-instant-1", "claude-instant-1.2"],
|
115 |
-
"Claude Instant",
|
116 |
-
"https://www.anthropic.com/index/introducing-claude",
|
117 |
-
"Claude Instant by Anthropic",
|
118 |
-
)
|
119 |
-
|
120 |
-
register_model_info(
|
121 |
-
["pplx-70b-online", "pplx-7b-online"],
|
122 |
-
"pplx-online-llms",
|
123 |
-
"https://blog.perplexity.ai/blog/introducing-pplx-online-llms",
|
124 |
-
"Online LLM API by Perplexity AI",
|
125 |
-
)
|
126 |
-
|
127 |
-
register_model_info(
|
128 |
-
["openhermes-2.5-mistral-7b"],
|
129 |
-
"OpenHermes-2.5-Mistral-7B",
|
130 |
-
"https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B",
|
131 |
-
"a mistral-based model fine-tuned on 1M GPT-4 outputs",
|
132 |
-
)
|
133 |
-
|
134 |
-
register_model_info(
|
135 |
-
["starling-lm-7b-alpha"],
|
136 |
-
"Starling-LM-7B-alpha",
|
137 |
-
"https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha",
|
138 |
-
"an open model trained using RLAIF by Berkeley",
|
139 |
-
)
|
140 |
-
|
141 |
-
register_model_info(
|
142 |
-
["tulu-2-dpo-70b"],
|
143 |
-
"Tulu 2",
|
144 |
-
"https://huggingface.co/allenai/tulu-2-dpo-70b",
|
145 |
-
"an instruction and RLHF model by UW/AllenAI",
|
146 |
-
)
|
147 |
-
|
148 |
-
register_model_info(
|
149 |
-
["yi-34b-chat", "yi-6b-chat"],
|
150 |
-
"Yi-Chat",
|
151 |
-
"https://huggingface.co/01-ai/Yi-34B-Chat",
|
152 |
-
"A large language model by 01 AI",
|
153 |
-
)
|
154 |
-
|
155 |
-
register_model_info(
|
156 |
-
["llama-2-70b-chat", "llama-2-34b-chat", "llama-2-13b-chat", "llama-2-7b-chat"],
|
157 |
-
"Llama 2",
|
158 |
-
"https://ai.meta.com/llama/",
|
159 |
-
"open foundation and fine-tuned chat models by Meta",
|
160 |
-
)
|
161 |
-
|
162 |
-
register_model_info(
|
163 |
-
[
|
164 |
-
"vicuna-33b",
|
165 |
-
"vicuna-33b-v1.3",
|
166 |
-
"vicuna-13b",
|
167 |
-
"vicuna-13b-v1.3",
|
168 |
-
"vicuna-7b",
|
169 |
-
"vicuna-7b-v1.3",
|
170 |
-
],
|
171 |
-
"Vicuna",
|
172 |
-
"https://lmsys.org/blog/2023-03-30-vicuna/",
|
173 |
-
"a chat assistant fine-tuned on user-shared conversations by LMSYS",
|
174 |
-
)
|
175 |
-
|
176 |
-
register_model_info(
|
177 |
-
["chatglm3-6b", "chatglm2-6b", "chatglm-6b"],
|
178 |
-
"ChatGLM",
|
179 |
-
"https://chatglm.cn/blog",
|
180 |
-
"an open bilingual dialogue language model by Tsinghua University",
|
181 |
-
)
|
182 |
-
|
183 |
-
register_model_info(
|
184 |
-
["openchat-3.5"],
|
185 |
-
"OpenChat 3.5",
|
186 |
-
"https://github.com/imoneoi/openchat",
|
187 |
-
"an open model fine-tuned on Mistral-7B using C-RLFT",
|
188 |
-
)
|
189 |
-
|
190 |
-
register_model_info(
|
191 |
-
["tenyxchat-7b-v1"],
|
192 |
-
"TenyxChat-7B",
|
193 |
-
"https://huggingface.co/tenyx/TenyxChat-7B-v1",
|
194 |
-
"an open model DPO trained on top of OpenChat-3.5 using Tenyx fine-tuning",
|
195 |
-
)
|
196 |
-
|
197 |
-
register_model_info(
|
198 |
-
["zephyr-7b-beta", "zephyr-7b-alpha"],
|
199 |
-
"Zephyr",
|
200 |
-
"https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha",
|
201 |
-
"a chatbot fine-tuned from Mistral by Hugging Face",
|
202 |
-
)
|
203 |
-
|
204 |
-
register_model_info(
|
205 |
-
["notus-7b-v1"],
|
206 |
-
"Notus",
|
207 |
-
"https://huggingface.co/argilla/notus-7b-v1",
|
208 |
-
"a chatbot fine-tuned from Zephyr SFT by Argilla",
|
209 |
-
)
|
210 |
-
|
211 |
-
register_model_info(
|
212 |
-
["catppt"],
|
213 |
-
"CatPPT",
|
214 |
-
"https://huggingface.co/rishiraj/CatPPT",
|
215 |
-
"a chatbot fine-tuned from a SLERP merged model by Rishiraj Acharya",
|
216 |
-
)
|
217 |
-
|
218 |
-
register_model_info(
|
219 |
-
["TinyLlama"],
|
220 |
-
"TinyLlama",
|
221 |
-
"https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
222 |
-
"The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.",
|
223 |
-
)
|
224 |
-
|
225 |
-
register_model_info(
|
226 |
-
["qwen-14b-chat"],
|
227 |
-
"Qwen",
|
228 |
-
"https://huggingface.co/Qwen/Qwen-14B-Chat",
|
229 |
-
"a large language model by Alibaba Cloud",
|
230 |
-
)
|
231 |
-
|
232 |
-
register_model_info(
|
233 |
-
["codellama-34b-instruct", "codellama-13b-instruct", "codellama-7b-instruct"],
|
234 |
-
"Code Llama",
|
235 |
-
"https://ai.meta.com/blog/code-llama-large-language-model-coding/",
|
236 |
-
"open foundation models for code by Meta",
|
237 |
-
)
|
238 |
-
|
239 |
-
register_model_info(
|
240 |
-
["wizardlm-70b", "wizardlm-30b", "wizardlm-13b"],
|
241 |
-
"WizardLM",
|
242 |
-
"https://github.com/nlpxucan/WizardLM",
|
243 |
-
"an instruction-following LLM using evol-instruct by Microsoft",
|
244 |
-
)
|
245 |
-
|
246 |
-
register_model_info(
|
247 |
-
["wizardcoder-15b-v1.0"],
|
248 |
-
"WizardLM",
|
249 |
-
"https://github.com/nlpxucan/WizardLM/tree/main/WizardCoder",
|
250 |
-
"Empowering Code Large Language Models with Evol-Instruct",
|
251 |
-
)
|
252 |
-
|
253 |
-
register_model_info(
|
254 |
-
["mpt-7b-chat", "mpt-30b-chat"],
|
255 |
-
"MPT-Chat",
|
256 |
-
"https://www.mosaicml.com/blog/mpt-30b",
|
257 |
-
"a chatbot fine-tuned from MPT by MosaicML",
|
258 |
-
)
|
259 |
-
|
260 |
-
register_model_info(
|
261 |
-
["guanaco-33b", "guanaco-65b"],
|
262 |
-
"Guanaco",
|
263 |
-
"https://github.com/artidoro/qlora",
|
264 |
-
"a model fine-tuned with QLoRA by UW",
|
265 |
-
)
|
266 |
-
|
267 |
-
register_model_info(
|
268 |
-
["gpt4all-13b-snoozy"],
|
269 |
-
"GPT4All-Snoozy",
|
270 |
-
"https://github.com/nomic-ai/gpt4all",
|
271 |
-
"a finetuned LLaMA model on assistant style data by Nomic AI",
|
272 |
-
)
|
273 |
-
|
274 |
-
register_model_info(
|
275 |
-
["koala-13b"],
|
276 |
-
"Koala",
|
277 |
-
"https://bair.berkeley.edu/blog/2023/04/03/koala",
|
278 |
-
"a dialogue model for academic research by BAIR",
|
279 |
-
)
|
280 |
-
|
281 |
-
register_model_info(
|
282 |
-
["RWKV-4-Raven-14B"],
|
283 |
-
"RWKV-4-Raven",
|
284 |
-
"https://huggingface.co/BlinkDL/rwkv-4-raven",
|
285 |
-
"an RNN with transformer-level LLM performance",
|
286 |
-
)
|
287 |
-
|
288 |
-
register_model_info(
|
289 |
-
["alpaca-13b"],
|
290 |
-
"Alpaca",
|
291 |
-
"https://crfm.stanford.edu/2023/03/13/alpaca.html",
|
292 |
-
"a model fine-tuned from LLaMA on instruction-following demonstrations by Stanford",
|
293 |
-
)
|
294 |
-
|
295 |
-
register_model_info(
|
296 |
-
["oasst-pythia-12b"],
|
297 |
-
"OpenAssistant (oasst)",
|
298 |
-
"https://open-assistant.io",
|
299 |
-
"an Open Assistant for everyone by LAION",
|
300 |
-
)
|
301 |
-
|
302 |
-
register_model_info(
|
303 |
-
["oasst-sft-7-llama-30b"],
|
304 |
-
"OpenAssistant (oasst)",
|
305 |
-
"https://open-assistant.io",
|
306 |
-
"an Open Assistant for everyone by LAION",
|
307 |
-
)
|
308 |
-
|
309 |
-
register_model_info(
|
310 |
-
["palm-2"],
|
311 |
-
"PaLM 2 Chat",
|
312 |
-
"https://cloud.google.com/vertex-ai/docs/release-notes#May_10_2023",
|
313 |
-
"PaLM 2 for Chat (chat-bison@001) by Google",
|
314 |
-
)
|
315 |
-
|
316 |
-
register_model_info(
|
317 |
-
["llama-7b", "llama-13b"],
|
318 |
-
"LLaMA",
|
319 |
-
"https://arxiv.org/abs/2302.13971",
|
320 |
-
"open and efficient foundation language models by Meta",
|
321 |
-
)
|
322 |
-
|
323 |
-
register_model_info(
|
324 |
-
["open-llama-7b-v2-open-instruct", "open-llama-7b-open-instruct"],
|
325 |
-
"Open LLaMa (Open Instruct)",
|
326 |
-
"https://medium.com/vmware-data-ml-blog/starter-llm-for-the-enterprise-instruction-tuning-openllama-7b-d05fc3bbaccc",
|
327 |
-
"Open LLaMa fine-tuned on instruction-following data by VMware",
|
328 |
-
)
|
329 |
-
|
330 |
-
register_model_info(
|
331 |
-
["dolly-v2-12b"],
|
332 |
-
"Dolly",
|
333 |
-
"https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm",
|
334 |
-
"an instruction-tuned open large language model by Databricks",
|
335 |
-
)
|
336 |
-
|
337 |
-
register_model_info(
|
338 |
-
["stablelm-tuned-alpha-7b"],
|
339 |
-
"StableLM",
|
340 |
-
"https://github.com/stability-AI/stableLM",
|
341 |
-
"Stability AI language models",
|
342 |
-
)
|
343 |
-
|
344 |
-
register_model_info(
|
345 |
-
["codet5p-6b"],
|
346 |
-
"CodeT5p-6b",
|
347 |
-
"https://huggingface.co/Salesforce/codet5p-6b",
|
348 |
-
"Code completion model released by Salesforce",
|
349 |
-
)
|
350 |
-
|
351 |
-
register_model_info(
|
352 |
-
["fastchat-t5-3b", "fastchat-t5-3b-v1.0"],
|
353 |
-
"FastChat-T5",
|
354 |
-
"https://huggingface.co/lmsys/fastchat-t5-3b-v1.0",
|
355 |
-
"a chat assistant fine-tuned from FLAN-T5 by LMSYS",
|
356 |
-
)
|
357 |
-
|
358 |
-
register_model_info(
|
359 |
-
["phoenix-inst-chat-7b"],
|
360 |
-
"Phoenix-7B",
|
361 |
-
"https://huggingface.co/FreedomIntelligence/phoenix-inst-chat-7b",
|
362 |
-
"a multilingual chat assistant fine-tuned from Bloomz to democratize ChatGPT across languages by CUHK(SZ)",
|
363 |
-
)
|
364 |
-
|
365 |
-
register_model_info(
|
366 |
-
["realm-7b-v1"],
|
367 |
-
"ReaLM",
|
368 |
-
"https://github.com/FreedomIntelligence/ReaLM",
|
369 |
-
"A chatbot fine-tuned from LLaMA2 with data generated via iterative calls to UserGPT and ChatGPT by CUHK(SZ) and SRIBD.",
|
370 |
-
)
|
371 |
-
|
372 |
-
register_model_info(
|
373 |
-
["billa-7b-sft"],
|
374 |
-
"BiLLa-7B-SFT",
|
375 |
-
"https://huggingface.co/Neutralzz/BiLLa-7B-SFT",
|
376 |
-
"an instruction-tuned bilingual LLaMA with enhanced reasoning ability by an independent researcher",
|
377 |
-
)
|
378 |
-
|
379 |
-
register_model_info(
|
380 |
-
["h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2"],
|
381 |
-
"h2oGPT-GM-7b",
|
382 |
-
"https://huggingface.co/h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2",
|
383 |
-
"an instruction-tuned OpenLLaMA with enhanced conversational ability by H2O.ai",
|
384 |
-
)
|
385 |
-
|
386 |
-
register_model_info(
|
387 |
-
["baize-v2-7b", "baize-v2-13b"],
|
388 |
-
"Baize v2",
|
389 |
-
"https://github.com/project-baize/baize-chatbot#v2",
|
390 |
-
"A chatbot fine-tuned from LLaMA with ChatGPT self-chat data and Self-Disillation with Feedback (SDF) by UCSD and SYSU.",
|
391 |
-
)
|
392 |
-
|
393 |
-
register_model_info(
|
394 |
-
[
|
395 |
-
"airoboros-l2-7b-2.1",
|
396 |
-
"airoboros-l2-13b-2.1",
|
397 |
-
"airoboros-c34b-2.1",
|
398 |
-
"airoboros-l2-70b-2.1",
|
399 |
-
],
|
400 |
-
"airoboros",
|
401 |
-
"https://huggingface.co/jondurbin/airoboros-l2-70b-2.1",
|
402 |
-
"an instruction-tuned LlaMa model tuned with 100% synthetic instruction-response pairs from GPT4",
|
403 |
-
)
|
404 |
-
|
405 |
-
register_model_info(
|
406 |
-
[
|
407 |
-
"spicyboros-7b-2.2",
|
408 |
-
"spicyboros-13b-2.2",
|
409 |
-
"spicyboros-70b-2.2",
|
410 |
-
],
|
411 |
-
"spicyboros",
|
412 |
-
"https://huggingface.co/jondurbin/spicyboros-70b-2.2",
|
413 |
-
"de-aligned versions of the airoboros models",
|
414 |
-
)
|
415 |
-
|
416 |
-
register_model_info(
|
417 |
-
["Robin-7b-v2", "Robin-13b-v2", "Robin-33b-v2"],
|
418 |
-
"Robin-v2",
|
419 |
-
"https://huggingface.co/OptimalScale/robin-7b-v2-delta",
|
420 |
-
"A chatbot fine-tuned from LLaMA-7b, achieving competitive performance on chitchat, commonsense reasoning and instruction-following tasks, by OptimalScale, HKUST.",
|
421 |
-
)
|
422 |
-
|
423 |
-
register_model_info(
|
424 |
-
["manticore-13b-chat"],
|
425 |
-
"Manticore 13B Chat",
|
426 |
-
"https://huggingface.co/openaccess-ai-collective/manticore-13b-chat-pyg",
|
427 |
-
"A chatbot fine-tuned from LlaMa across several CoT and chat datasets.",
|
428 |
-
)
|
429 |
-
|
430 |
-
register_model_info(
|
431 |
-
["redpajama-incite-7b-chat"],
|
432 |
-
"RedPajama-INCITE-7B-Chat",
|
433 |
-
"https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Chat",
|
434 |
-
"A chatbot fine-tuned from RedPajama-INCITE-7B-Base by Together",
|
435 |
-
)
|
436 |
-
|
437 |
-
register_model_info(
|
438 |
-
[
|
439 |
-
"falcon-7b",
|
440 |
-
"falcon-7b-instruct",
|
441 |
-
"falcon-40b",
|
442 |
-
"falcon-40b-instruct",
|
443 |
-
"falcon-180b",
|
444 |
-
"falcon-180b-chat",
|
445 |
-
],
|
446 |
-
"Falcon",
|
447 |
-
"https://huggingface.co/tiiuae/falcon-180B",
|
448 |
-
"TII's flagship series of large language models",
|
449 |
-
)
|
450 |
-
|
451 |
-
register_model_info(
|
452 |
-
["tigerbot-7b-sft"],
|
453 |
-
"Tigerbot",
|
454 |
-
"https://huggingface.co/TigerResearch/tigerbot-7b-sft",
|
455 |
-
"TigerBot is a large-scale language model (LLM) with multiple languages and tasks.",
|
456 |
-
)
|
457 |
-
|
458 |
-
register_model_info(
|
459 |
-
["internlm-chat-7b", "internlm-chat-7b-8k"],
|
460 |
-
"InternLM",
|
461 |
-
"https://huggingface.co/internlm/internlm-chat-7b",
|
462 |
-
"InternLM is a multi-language large-scale language model (LLM), developed by SHLAB.",
|
463 |
-
)
|
464 |
-
|
465 |
-
register_model_info(
|
466 |
-
["Qwen-7B-Chat"],
|
467 |
-
"Qwen",
|
468 |
-
"https://huggingface.co/Qwen/Qwen-7B-Chat",
|
469 |
-
"Qwen is a multi-language large-scale language model (LLM), developed by Damo Academy.",
|
470 |
-
)
|
471 |
-
|
472 |
-
register_model_info(
|
473 |
-
["Llama2-Chinese-13b-Chat", "LLama2-Chinese-13B"],
|
474 |
-
"Llama2-Chinese",
|
475 |
-
"https://huggingface.co/FlagAlpha/Llama2-Chinese-13b-Chat",
|
476 |
-
"Llama2-Chinese is a multi-language large-scale language model (LLM), developed by FlagAlpha.",
|
477 |
-
)
|
478 |
-
|
479 |
-
register_model_info(
|
480 |
-
["Chinese-Alpaca-2-7B", "Chinese-Alpaca-2-13B"],
|
481 |
-
"Chinese-Alpaca",
|
482 |
-
"https://huggingface.co/hfl/chinese-alpaca-2-13b",
|
483 |
-
"New extended Chinese vocabulary beyond Llama-2, open-sourcing the Chinese LLaMA-2 and Alpaca-2 LLMs.",
|
484 |
-
)
|
485 |
-
|
486 |
-
register_model_info(
|
487 |
-
["Vigogne-2-7B-Instruct", "Vigogne-2-13B-Instruct"],
|
488 |
-
"Vigogne-Instruct",
|
489 |
-
"https://huggingface.co/bofenghuang/vigogne-2-7b-instruct",
|
490 |
-
"Vigogne-Instruct is a French large language model (LLM) optimized for instruction-following, developed by Bofeng Huang",
|
491 |
-
)
|
492 |
-
|
493 |
-
register_model_info(
|
494 |
-
["Vigogne-2-7B-Chat", "Vigogne-2-13B-Chat"],
|
495 |
-
"Vigogne-Chat",
|
496 |
-
"https://huggingface.co/bofenghuang/vigogne-2-7b-chat",
|
497 |
-
"Vigogne-Chat is a French large language model (LLM) optimized for instruction-following and multi-turn dialogues, developed by Bofeng Huang",
|
498 |
-
)
|
499 |
-
|
500 |
-
register_model_info(
|
501 |
-
["stable-vicuna-13B-HF"],
|
502 |
-
"stable-vicuna",
|
503 |
-
"https://huggingface.co/TheBloke/stable-vicuna-13B-HF",
|
504 |
-
"StableVicuna is a Vicuna model fine-tuned using RLHF via PPO on various conversational and instructional datasets.",
|
505 |
-
)
|
506 |
-
|
507 |
-
register_model_info(
|
508 |
-
["deluxe-chat-v1", "deluxe-chat-v1.1", "deluxe-chat-v1.2"],
|
509 |
-
"DeluxeChat",
|
510 |
-
"",
|
511 |
-
"Deluxe Chat",
|
512 |
-
)
|
513 |
-
|
514 |
-
register_model_info(
|
515 |
-
[
|
516 |
-
"Xwin-LM-7B-V0.1",
|
517 |
-
"Xwin-LM-13B-V0.1",
|
518 |
-
"Xwin-LM-70B-V0.1",
|
519 |
-
"Xwin-LM-7B-V0.2",
|
520 |
-
"Xwin-LM-13B-V0.2",
|
521 |
-
],
|
522 |
-
"Xwin-LM",
|
523 |
-
"https://github.com/Xwin-LM/Xwin-LM",
|
524 |
-
"Chat models developed by Xwin-LM team",
|
525 |
-
)
|
526 |
-
|
527 |
-
register_model_info(
|
528 |
-
["lemur-70b-chat"],
|
529 |
-
"Lemur-Chat",
|
530 |
-
"https://huggingface.co/OpenLemur/lemur-70b-chat-v1",
|
531 |
-
"an openly accessible language model optimized for both natural language and coding capabilities ",
|
532 |
-
)
|
533 |
-
|
534 |
-
register_model_info(
|
535 |
-
["Mistral-7B-OpenOrca"],
|
536 |
-
"Open-Orca",
|
537 |
-
"https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca",
|
538 |
-
"A fine-tune of [Mistral 7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) using [OpenOrca dataset](https://huggingface.co/datasets/Open-Orca/OpenOrca)",
|
539 |
-
)
|
540 |
-
|
541 |
-
register_model_info(
|
542 |
-
["dolphin-2.2.1-mistral-7b"],
|
543 |
-
"dolphin-mistral",
|
544 |
-
"https://huggingface.co/ehartford/dolphin-2.2.1-mistral-7b",
|
545 |
-
"An uncensored fine-tuned Mistral 7B",
|
546 |
-
)
|
547 |
-
|
548 |
-
register_model_info(
|
549 |
-
[
|
550 |
-
"AquilaChat-7B",
|
551 |
-
"AquilaChat2-7B",
|
552 |
-
"AquilaChat2-34B",
|
553 |
-
],
|
554 |
-
"Aquila-Chat",
|
555 |
-
"https://huggingface.co/BAAI/AquilaChat2-34B",
|
556 |
-
"Chat models developed by BAAI team",
|
557 |
-
)
|
558 |
-
|
559 |
-
register_model_info(
|
560 |
-
["xDAN-L1-Chat-RL-v1"],
|
561 |
-
"xDAN-L1-Chat",
|
562 |
-
"https://huggingface.co/xDAN-AI/xDAN-L1-Chat-RL-v1",
|
563 |
-
"A large language chat model created by xDAN-AI.",
|
564 |
-
)
|
565 |
-
|
566 |
-
register_model_info(
|
567 |
-
["MetaMath-70B-V1.0", "MetaMath-7B-V1.0"],
|
568 |
-
"MetaMath",
|
569 |
-
"https://huggingface.co/meta-math",
|
570 |
-
"MetaMath is a finetune of Llama2 on [MetaMathQA](https://huggingface.co/datasets/meta-math/MetaMathQA) that specializes in mathematical reasoning.",
|
571 |
-
)
|
572 |
-
|
573 |
-
register_model_info(
|
574 |
-
["Yuan2-2B-hf", "Yuan2-51B-hf", "Yuan2-102B-hf"],
|
575 |
-
"IEIYuan",
|
576 |
-
"https://huggingface.co/IEITYuan",
|
577 |
-
"Yuan2 is a Basemodel developed by IEI.",
|
578 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
arena_elo/elo_rating/utils.py
CHANGED
@@ -3,12 +3,20 @@ import pytz
|
|
3 |
import PIL
|
4 |
import os
|
5 |
|
|
|
|
|
|
|
|
|
6 |
def detect_language(text: str) -> str:
|
7 |
"""Detect the langauge of a string."""
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
|
|
|
|
|
|
|
|
12 |
|
13 |
polyglot_logger.setLevel("ERROR")
|
14 |
|
|
|
3 |
import PIL
|
4 |
import os
|
5 |
|
6 |
+
import sys
|
7 |
+
sys.path.append('../')
|
8 |
+
from model.model_registry import get_model_info
|
9 |
+
|
10 |
def detect_language(text: str) -> str:
|
11 |
"""Detect the langauge of a string."""
|
12 |
+
try:
|
13 |
+
import polyglot # pip3 install polyglot pyicu pycld2
|
14 |
+
from polyglot.detect import Detector
|
15 |
+
from polyglot.detect.base import logger as polyglot_logger
|
16 |
+
import pycld2
|
17 |
+
except ImportError as e:
|
18 |
+
print("Please install the required libraries: polyglot, pycld2: pip3 install polyglot pyicu pycld2")
|
19 |
+
exit(1)
|
20 |
|
21 |
polyglot_logger.setLevel("ERROR")
|
22 |
|
arena_elo/results/20240809/elo_results_image_editing.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe056dd818906f7d714c47fdba95fdcd01db8b8aa8490cfa0399cf2baab086e0
|
3 |
+
size 63251
|
arena_elo/results/20240809/image_editing_leaderboard.csv
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
-
MagicBrush,MagicBrush,1112.4518849527376,1116.3228088792275,CC-BY-4.0,"The Ohio State University, University of Waterloo",https://osu-nlp-group.github.io/MagicBrush
|
3 |
-
InfEdit,InfEdit,1077.2249767545911,1077.0861206067664,CC BY-NC-ND 4.0,"University of Michigan, University of California, Berkeley",https://
|
4 |
-
CosXLEdit,CosXLEdit,1064.682441262532,1065.599445502775,cosxl-nc-community,Stability AI,https://huggingface.co/
|
5 |
InstructPix2Pix,InstructPix2Pix,1035.4054226880567,1033.2086681336118,"Copyright 2023 Timothy Brooks, Aleksander Holynski, Alexei A. Efros","University of California, Berkeley",https://www.timothybrooks.com/instruct-pix2pix
|
6 |
PNP,PNP,1001.4916130267786,1006.0747389504991,-,Weizmann Institute of Science,https://github.com/MichalGeyer/plug-and-play
|
7 |
-
Prompt2prompt,Prompt2prompt,990.9755779591587,992.0796321838297,Apache-2.0,"Google, Tel Aviv University",https://prompt-to-prompt.github.io
|
8 |
-
CycleDiffusion,CycleDiffusion,938.5637450408492,931.6722574678079,X11,Carnegie Mellon University,https://github.com/ChenWu98/cycle-diffusion
|
9 |
SDEdit,SDEdit,925.0466906054369,923.4126211812364,MIT License,Stanford University,https://sde-image-editing.github.io
|
10 |
-
Pix2PixZero,Pix2PixZero,854.1576477098594,854.5437070942457,MIT License,"Carnegie Mellon University, Adobe Research",https://pix2pixzero.github.io
|
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
+
MagicBrush,MagicBrush,1112.4518849527376,1116.3228088792275,CC-BY-4.0,"The Ohio State University, University of Waterloo",https://osu-nlp-group.github.io/MagicBrush/
|
3 |
+
InfEdit,InfEdit,1077.2249767545911,1077.0861206067664,CC BY-NC-ND 4.0,"University of Michigan, University of California, Berkeley",https://sled-group.github.io/InfEdit/
|
4 |
+
CosXLEdit,CosXLEdit,1064.682441262532,1065.599445502775,cosxl-nc-community,Stability AI,https://huggingface.co/stabilityai/cosxl
|
5 |
InstructPix2Pix,InstructPix2Pix,1035.4054226880567,1033.2086681336118,"Copyright 2023 Timothy Brooks, Aleksander Holynski, Alexei A. Efros","University of California, Berkeley",https://www.timothybrooks.com/instruct-pix2pix
|
6 |
PNP,PNP,1001.4916130267786,1006.0747389504991,-,Weizmann Institute of Science,https://github.com/MichalGeyer/plug-and-play
|
7 |
+
Prompt2prompt,Prompt2prompt,990.9755779591587,992.0796321838297,Apache-2.0,"Google, Tel Aviv University",https://prompt-to-prompt.github.io/
|
8 |
+
CycleDiffusion,CycleDiffusion,938.5637450408492,931.6722574678079,X11,Carnegie Mellon University,https://github.com/ChenWu98/cycle-diffusion?tab=readme-ov-file
|
9 |
SDEdit,SDEdit,925.0466906054369,923.4126211812364,MIT License,Stanford University,https://sde-image-editing.github.io
|
10 |
+
Pix2PixZero,Pix2PixZero,854.1576477098594,854.5437070942457,MIT License,"Carnegie Mellon University, Adobe Research",https://pix2pixzero.github.io/
|
arena_elo/results/20240819/clean_battle_video_generation.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
arena_elo/results/20240819/elo_results_video_generation.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c42587eea0a948a9cbd5cca3b94e8141e70bcb9a0ae86646bc22973ea44332a
|
3 |
+
size 66377
|
arena_elo/results/20240819/video_generation_leaderboard.csv
CHANGED
@@ -1,11 +1,11 @@
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
-
StableVideoDiffusion,StableVideoDiffusion,1143.2191298083312,1147.5193110801638,
|
3 |
-
|
4 |
VideoCrafter2,VideoCrafter2,1076.7797371219958,1079.3282160297892,Apache 2.0,Tencent AI Lab,https://ailab-cvc.github.io/videocrafter2/
|
5 |
AnimateDiff,AnimateDiff,1071.9888858847867,1073.0390578256142,creativeml-openrail-m,"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",https://fal.ai/models/fast-animatediff-t2v
|
6 |
-
CogVideoX,CogVideoX,1046.417835182279,1048.070311411077,
|
7 |
LaVie,LaVie,999.6270938177271,1001.6816790561326,Apache 2.0,Shanghai AI Lab,https://github.com/Vchitect/LaVie
|
8 |
-
|
9 |
OpenSora,OpenSora,916.0498767986071,917.9961732110512,Apache 2.0,HPC-AI Tech,https://github.com/hpcaitech/Open-Sora
|
10 |
ModelScope,ModelScope,865.3398542588171,868.9309810684535,cc-by-nc-4.0,Alibaba Group,https://arxiv.org/abs/2308.06571
|
11 |
-
|
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
+
StableVideoDiffusion,StableVideoDiffusion,1143.2191298083312,1147.5193110801638,SVD-nc-community,Stability AI,https://fal.ai/models/fal-ai/fast-svd/text-to-video/api
|
3 |
+
T2V-Turbo,T2V-Turbo,1092.4551860072625,1093.43771637836,cc-by-nc-4.0,"University of California, Santa Barbara",https://github.com/Ji4chenLi/t2v-turbo
|
4 |
VideoCrafter2,VideoCrafter2,1076.7797371219958,1079.3282160297892,Apache 2.0,Tencent AI Lab,https://ailab-cvc.github.io/videocrafter2/
|
5 |
AnimateDiff,AnimateDiff,1071.9888858847867,1073.0390578256142,creativeml-openrail-m,"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",https://fal.ai/models/fast-animatediff-t2v
|
6 |
+
CogVideoX,CogVideoX,1046.417835182279,1048.070311411077,CogVideoX LICENSE,THUDM,https://github.com/THUDM/CogVideo
|
7 |
LaVie,LaVie,999.6270938177271,1001.6816790561326,Apache 2.0,Shanghai AI Lab,https://github.com/Vchitect/LaVie
|
8 |
+
OpenSora v1.2,OpenSora v1.2,924.0203683794037,903.7447677401199,Apache 2.0,HPC-AI Tech,https://github.com/hpcaitech/Open-Sora
|
9 |
OpenSora,OpenSora,916.0498767986071,917.9961732110512,Apache 2.0,HPC-AI Tech,https://github.com/hpcaitech/Open-Sora
|
10 |
ModelScope,ModelScope,865.3398542588171,868.9309810684535,cc-by-nc-4.0,Alibaba Group,https://arxiv.org/abs/2308.06571
|
11 |
+
AnimateDiff Turbo,AnimateDiff Turbo,864.1020327407906,866.2517861992403,creativeml-openrail-m,"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",https://fal.ai/models/fast-animatediff-t2v-turbo
|
arena_elo/results/20240820/clean_battle_t2i_generation.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
arena_elo/results/20240820/elo_results_t2i_generation.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a55e58f091b88f7cc16382c56b7d655cbcd86bd23c3f3de214ef5484edfd1e4d
|
3 |
+
size 88247
|
arena_elo/results/20240820/t2i_generation_leaderboard.csv
CHANGED
@@ -1,18 +1,18 @@
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
-
PlayGround V2.5,PlayGround V2.5,
|
3 |
-
|
4 |
-
PlayGround V2,PlayGround V2,
|
5 |
-
|
6 |
-
HunyuanDiT,HunyuanDiT,
|
7 |
-
StableCascade,StableCascade,
|
8 |
-
AuraFlow,AuraFlow,1028.
|
9 |
-
PixArtAlpha,PixArtAlpha,1027.
|
10 |
-
|
11 |
-
PixArtSigma,PixArtSigma,1022.
|
12 |
-
SD3,SD3,992.
|
13 |
-
Kolors,Kolors,980.
|
14 |
-
SDXL,SDXL,968.
|
15 |
-
SDXLTurbo,SDXLTurbo,
|
16 |
-
LCM(v1.5/XL),LCM(v1.5/XL),909.
|
17 |
-
OpenJourney,OpenJourney,
|
18 |
-
LCM,LCM,
|
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
+
PlayGround V2.5,PlayGround V2.5,1132.890132362509,1133.2132994392719,Playground v2.5 Community License,Playground,https://huggingface.co/playgroundai/playground-v2.5-1024px-aesthetic
|
3 |
+
FLUX.1-dev,FLUX.1-dev,1130.7155148553181,1147.438205902597,flux-1-dev-non-commercial-license (other),Black Forest Labs,https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux
|
4 |
+
PlayGround V2,PlayGround V2,1076.980283311089,1074.3249013429554,Playground v2 Community License,Playground,https://huggingface.co/playgroundai/playground-v2-1024px-aesthetic
|
5 |
+
FLUX.1-schnell,FLUX.1-schnell,1066.4020402818517,1062.8730931830312,flux-1-dev-non-commercial-license (other),Black Forest Labs,https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux
|
6 |
+
HunyuanDiT,HunyuanDiT,1053.572229850288,1038.703774370632,tencent-hunyuan-community,Tencent,https://github.com/Tencent/HunyuanDiT
|
7 |
+
StableCascade,StableCascade,1040.7649360610326,1042.7543257183775,stable-cascade-nc-community (other),Stability AI,https://fal.ai/models/stable-cascade/api
|
8 |
+
AuraFlow,AuraFlow,1028.6476636951288,1021.8888350731722,Apache-2.0,Fal.AI,https://huggingface.co/fal/AuraFlow
|
9 |
+
PixArtAlpha,PixArtAlpha,1027.5984425989943,1016.2046020683781,openrail++,PixArt-alpha,https://huggingface.co/PixArt-alpha/PixArt-XL-2-1024-MS
|
10 |
+
SDXL-Lightning,SDXL-Lightning,1027.28056074664,1030.1892737463068,openrail++,ByteDance,https://huggingface.co/ByteDance/SDXL-Lightning
|
11 |
+
PixArtSigma,PixArtSigma,1022.278191734517,1020.5244848333533,openrail++,PixArt-alpha,https://github.com/PixArt-alpha/PixArt-sigma
|
12 |
+
SD3,SD3,992.4683734086498,986.8567562452171,stabilityai-nc-research-community,Stability AI,https://huggingface.co/blog/sd3
|
13 |
+
Kolors,Kolors,980.5720834009985,977.940378018964,Apache-2.0,Kwai Kolors,https://huggingface.co/Kwai-Kolors/Kolors
|
14 |
+
SDXL,SDXL,968.286211097592,967.3440457396239,openrail++,Stability AI,https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0
|
15 |
+
SDXLTurbo,SDXLTurbo,916.9664722660239,912.8537618702409,sai-nc-community (other),Stability AI,https://huggingface.co/stabilityai/sdxl-turbo
|
16 |
+
LCM(v1.5/XL),LCM(v1.5/XL),909.2587727377047,901.7564308088253,openrail++,Latent Consistency,https://fal.ai/models/fast-lcm-diffusion-turbo
|
17 |
+
OpenJourney,OpenJourney,831.6663894963693,825.0727531727863,creativeml-openrail-m,PromptHero,https://huggingface.co/prompthero/openjourney
|
18 |
+
LCM,LCM,793.6517020952907,806.0428762517084,MIT License,Tsinghua University,https://huggingface.co/SimianLuo/LCM_Dreamshaper_v7
|
arena_elo/results/latest/clean_battle_t2i_generation.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
arena_elo/results/latest/clean_battle_video_generation.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
arena_elo/results/latest/elo_results_image_editing.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe056dd818906f7d714c47fdba95fdcd01db8b8aa8490cfa0399cf2baab086e0
|
3 |
+
size 63251
|
arena_elo/results/latest/elo_results_t2i_generation.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a55e58f091b88f7cc16382c56b7d655cbcd86bd23c3f3de214ef5484edfd1e4d
|
3 |
+
size 88247
|
arena_elo/results/latest/elo_results_video_generation.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c42587eea0a948a9cbd5cca3b94e8141e70bcb9a0ae86646bc22973ea44332a
|
3 |
+
size 66377
|
arena_elo/results/latest/image_editing_leaderboard.csv
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
-
MagicBrush,MagicBrush,1112.4518849527376,1116.3228088792275,CC-BY-4.0,"The Ohio State University, University of Waterloo",https://osu-nlp-group.github.io/MagicBrush
|
3 |
-
InfEdit,InfEdit,1077.2249767545911,1077.0861206067664,CC BY-NC-ND 4.0,"University of Michigan, University of California, Berkeley",https://
|
4 |
-
CosXLEdit,CosXLEdit,1064.682441262532,1065.599445502775,cosxl-nc-community,Stability AI,https://huggingface.co/
|
5 |
InstructPix2Pix,InstructPix2Pix,1035.4054226880567,1033.2086681336118,"Copyright 2023 Timothy Brooks, Aleksander Holynski, Alexei A. Efros","University of California, Berkeley",https://www.timothybrooks.com/instruct-pix2pix
|
6 |
PNP,PNP,1001.4916130267786,1006.0747389504991,-,Weizmann Institute of Science,https://github.com/MichalGeyer/plug-and-play
|
7 |
-
Prompt2prompt,Prompt2prompt,990.9755779591587,992.0796321838297,Apache-2.0,"Google, Tel Aviv University",https://prompt-to-prompt.github.io
|
8 |
-
CycleDiffusion,CycleDiffusion,938.5637450408492,931.6722574678079,X11,Carnegie Mellon University,https://github.com/ChenWu98/cycle-diffusion
|
9 |
SDEdit,SDEdit,925.0466906054369,923.4126211812364,MIT License,Stanford University,https://sde-image-editing.github.io
|
10 |
-
Pix2PixZero,Pix2PixZero,854.1576477098594,854.5437070942457,MIT License,"Carnegie Mellon University, Adobe Research",https://pix2pixzero.github.io
|
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
+
MagicBrush,MagicBrush,1112.4518849527376,1116.3228088792275,CC-BY-4.0,"The Ohio State University, University of Waterloo",https://osu-nlp-group.github.io/MagicBrush/
|
3 |
+
InfEdit,InfEdit,1077.2249767545911,1077.0861206067664,CC BY-NC-ND 4.0,"University of Michigan, University of California, Berkeley",https://sled-group.github.io/InfEdit/
|
4 |
+
CosXLEdit,CosXLEdit,1064.682441262532,1065.599445502775,cosxl-nc-community,Stability AI,https://huggingface.co/stabilityai/cosxl
|
5 |
InstructPix2Pix,InstructPix2Pix,1035.4054226880567,1033.2086681336118,"Copyright 2023 Timothy Brooks, Aleksander Holynski, Alexei A. Efros","University of California, Berkeley",https://www.timothybrooks.com/instruct-pix2pix
|
6 |
PNP,PNP,1001.4916130267786,1006.0747389504991,-,Weizmann Institute of Science,https://github.com/MichalGeyer/plug-and-play
|
7 |
+
Prompt2prompt,Prompt2prompt,990.9755779591587,992.0796321838297,Apache-2.0,"Google, Tel Aviv University",https://prompt-to-prompt.github.io/
|
8 |
+
CycleDiffusion,CycleDiffusion,938.5637450408492,931.6722574678079,X11,Carnegie Mellon University,https://github.com/ChenWu98/cycle-diffusion?tab=readme-ov-file
|
9 |
SDEdit,SDEdit,925.0466906054369,923.4126211812364,MIT License,Stanford University,https://sde-image-editing.github.io
|
10 |
+
Pix2PixZero,Pix2PixZero,854.1576477098594,854.5437070942457,MIT License,"Carnegie Mellon University, Adobe Research",https://pix2pixzero.github.io/
|
arena_elo/results/latest/t2i_generation_leaderboard.csv
CHANGED
@@ -1,18 +1,18 @@
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
-
PlayGround V2.5,PlayGround V2.5,
|
3 |
-
|
4 |
-
PlayGround V2,PlayGround V2,
|
5 |
-
|
6 |
-
HunyuanDiT,HunyuanDiT,
|
7 |
-
StableCascade,StableCascade,
|
8 |
-
AuraFlow,AuraFlow,1028.
|
9 |
-
PixArtAlpha,PixArtAlpha,1027.
|
10 |
-
|
11 |
-
PixArtSigma,PixArtSigma,1022.
|
12 |
-
SD3,SD3,992.
|
13 |
-
Kolors,Kolors,980.
|
14 |
-
SDXL,SDXL,968.
|
15 |
-
SDXLTurbo,SDXLTurbo,
|
16 |
-
LCM(v1.5/XL),LCM(v1.5/XL),909.
|
17 |
-
OpenJourney,OpenJourney,
|
18 |
-
LCM,LCM,
|
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
+
PlayGround V2.5,PlayGround V2.5,1132.890132362509,1133.2132994392719,Playground v2.5 Community License,Playground,https://huggingface.co/playgroundai/playground-v2.5-1024px-aesthetic
|
3 |
+
FLUX.1-dev,FLUX.1-dev,1130.7155148553181,1147.438205902597,flux-1-dev-non-commercial-license (other),Black Forest Labs,https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux
|
4 |
+
PlayGround V2,PlayGround V2,1076.980283311089,1074.3249013429554,Playground v2 Community License,Playground,https://huggingface.co/playgroundai/playground-v2-1024px-aesthetic
|
5 |
+
FLUX.1-schnell,FLUX.1-schnell,1066.4020402818517,1062.8730931830312,flux-1-dev-non-commercial-license (other),Black Forest Labs,https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux
|
6 |
+
HunyuanDiT,HunyuanDiT,1053.572229850288,1038.703774370632,tencent-hunyuan-community,Tencent,https://github.com/Tencent/HunyuanDiT
|
7 |
+
StableCascade,StableCascade,1040.7649360610326,1042.7543257183775,stable-cascade-nc-community (other),Stability AI,https://fal.ai/models/stable-cascade/api
|
8 |
+
AuraFlow,AuraFlow,1028.6476636951288,1021.8888350731722,Apache-2.0,Fal.AI,https://huggingface.co/fal/AuraFlow
|
9 |
+
PixArtAlpha,PixArtAlpha,1027.5984425989943,1016.2046020683781,openrail++,PixArt-alpha,https://huggingface.co/PixArt-alpha/PixArt-XL-2-1024-MS
|
10 |
+
SDXL-Lightning,SDXL-Lightning,1027.28056074664,1030.1892737463068,openrail++,ByteDance,https://huggingface.co/ByteDance/SDXL-Lightning
|
11 |
+
PixArtSigma,PixArtSigma,1022.278191734517,1020.5244848333533,openrail++,PixArt-alpha,https://github.com/PixArt-alpha/PixArt-sigma
|
12 |
+
SD3,SD3,992.4683734086498,986.8567562452171,stabilityai-nc-research-community,Stability AI,https://huggingface.co/blog/sd3
|
13 |
+
Kolors,Kolors,980.5720834009985,977.940378018964,Apache-2.0,Kwai Kolors,https://huggingface.co/Kwai-Kolors/Kolors
|
14 |
+
SDXL,SDXL,968.286211097592,967.3440457396239,openrail++,Stability AI,https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0
|
15 |
+
SDXLTurbo,SDXLTurbo,916.9664722660239,912.8537618702409,sai-nc-community (other),Stability AI,https://huggingface.co/stabilityai/sdxl-turbo
|
16 |
+
LCM(v1.5/XL),LCM(v1.5/XL),909.2587727377047,901.7564308088253,openrail++,Latent Consistency,https://fal.ai/models/fast-lcm-diffusion-turbo
|
17 |
+
OpenJourney,OpenJourney,831.6663894963693,825.0727531727863,creativeml-openrail-m,PromptHero,https://huggingface.co/prompthero/openjourney
|
18 |
+
LCM,LCM,793.6517020952907,806.0428762517084,MIT License,Tsinghua University,https://huggingface.co/SimianLuo/LCM_Dreamshaper_v7
|
arena_elo/results/latest/video_generation_leaderboard.csv
CHANGED
@@ -1,11 +1,11 @@
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
StableVideoDiffusion,StableVideoDiffusion,1143.2191298083312,1147.5193110801638,SVD-nc-community,Stability AI,https://fal.ai/models/fal-ai/fast-svd/text-to-video/api
|
3 |
-
|
4 |
VideoCrafter2,VideoCrafter2,1076.7797371219958,1079.3282160297892,Apache 2.0,Tencent AI Lab,https://ailab-cvc.github.io/videocrafter2/
|
5 |
AnimateDiff,AnimateDiff,1071.9888858847867,1073.0390578256142,creativeml-openrail-m,"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",https://fal.ai/models/fast-animatediff-t2v
|
6 |
-
CogVideoX,CogVideoX,1046.417835182279,1048.070311411077,CogVideoX
|
7 |
LaVie,LaVie,999.6270938177271,1001.6816790561326,Apache 2.0,Shanghai AI Lab,https://github.com/Vchitect/LaVie
|
8 |
-
|
9 |
OpenSora,OpenSora,916.0498767986071,917.9961732110512,Apache 2.0,HPC-AI Tech,https://github.com/hpcaitech/Open-Sora
|
10 |
ModelScope,ModelScope,865.3398542588171,868.9309810684535,cc-by-nc-4.0,Alibaba Group,https://arxiv.org/abs/2308.06571
|
11 |
-
|
|
|
1 |
key,Model,Arena Elo rating (anony),Arena Elo rating (full),License,Organization,Link
|
2 |
StableVideoDiffusion,StableVideoDiffusion,1143.2191298083312,1147.5193110801638,SVD-nc-community,Stability AI,https://fal.ai/models/fal-ai/fast-svd/text-to-video/api
|
3 |
+
T2V-Turbo,T2V-Turbo,1092.4551860072625,1093.43771637836,cc-by-nc-4.0,"University of California, Santa Barbara",https://github.com/Ji4chenLi/t2v-turbo
|
4 |
VideoCrafter2,VideoCrafter2,1076.7797371219958,1079.3282160297892,Apache 2.0,Tencent AI Lab,https://ailab-cvc.github.io/videocrafter2/
|
5 |
AnimateDiff,AnimateDiff,1071.9888858847867,1073.0390578256142,creativeml-openrail-m,"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",https://fal.ai/models/fast-animatediff-t2v
|
6 |
+
CogVideoX,CogVideoX,1046.417835182279,1048.070311411077,CogVideoX LICENSE,THUDM,https://github.com/THUDM/CogVideo
|
7 |
LaVie,LaVie,999.6270938177271,1001.6816790561326,Apache 2.0,Shanghai AI Lab,https://github.com/Vchitect/LaVie
|
8 |
+
OpenSora v1.2,OpenSora v1.2,924.0203683794037,903.7447677401199,Apache 2.0,HPC-AI Tech,https://github.com/hpcaitech/Open-Sora
|
9 |
OpenSora,OpenSora,916.0498767986071,917.9961732110512,Apache 2.0,HPC-AI Tech,https://github.com/hpcaitech/Open-Sora
|
10 |
ModelScope,ModelScope,865.3398542588171,868.9309810684535,cc-by-nc-4.0,Alibaba Group,https://arxiv.org/abs/2308.06571
|
11 |
+
AnimateDiff Turbo,AnimateDiff Turbo,864.1020327407906,866.2517861992403,creativeml-openrail-m,"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",https://fal.ai/models/fast-animatediff-t2v-turbo
|
arena_elo/update_elo_rating.sh
CHANGED
@@ -48,17 +48,14 @@ mv ./elo_results_$video_generation_battle_cutoff_date.pkl ./results/$video_gener
|
|
48 |
# generat the leaderboard
|
49 |
|
50 |
python -m elo_rating.generate_leaderboard \
|
51 |
-
--model_info_file "./edition_model_info.json" \
|
52 |
--elo_rating_pkl "./results/$edition_battle_cutoff_date/elo_results_image_editing.pkl" \
|
53 |
--output_csv "./results/$edition_battle_cutoff_date/image_editing_leaderboard.csv"
|
54 |
|
55 |
python -m elo_rating.generate_leaderboard \
|
56 |
-
--model_info_file "./generation_model_info.json" \
|
57 |
--elo_rating_pkl "./results/$generation_battle_cutoff_date/elo_results_t2i_generation.pkl" \
|
58 |
--output_csv "./results/$generation_battle_cutoff_date/t2i_generation_leaderboard.csv"
|
59 |
|
60 |
python -m elo_rating.generate_leaderboard \
|
61 |
-
--model_info_file "./video_generation_model_info.json" \
|
62 |
--elo_rating_pkl "./results/$video_generation_battle_cutoff_date/elo_results_video_generation.pkl" \
|
63 |
--output_csv "./results/$video_generation_battle_cutoff_date/video_generation_leaderboard.csv"
|
64 |
|
|
|
48 |
# generat the leaderboard
|
49 |
|
50 |
python -m elo_rating.generate_leaderboard \
|
|
|
51 |
--elo_rating_pkl "./results/$edition_battle_cutoff_date/elo_results_image_editing.pkl" \
|
52 |
--output_csv "./results/$edition_battle_cutoff_date/image_editing_leaderboard.csv"
|
53 |
|
54 |
python -m elo_rating.generate_leaderboard \
|
|
|
55 |
--elo_rating_pkl "./results/$generation_battle_cutoff_date/elo_results_t2i_generation.pkl" \
|
56 |
--output_csv "./results/$generation_battle_cutoff_date/t2i_generation_leaderboard.csv"
|
57 |
|
58 |
python -m elo_rating.generate_leaderboard \
|
|
|
59 |
--elo_rating_pkl "./results/$video_generation_battle_cutoff_date/elo_results_video_generation.pkl" \
|
60 |
--output_csv "./results/$video_generation_battle_cutoff_date/video_generation_leaderboard.csv"
|
61 |
|
model/model_registry.py
CHANGED
@@ -1,16 +1,42 @@
|
|
1 |
from collections import namedtuple
|
2 |
from typing import List
|
3 |
|
4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
model_info = {}
|
6 |
|
7 |
def register_model_info(
|
8 |
-
full_names: List[str], simple_name: str, link: str, description: str
|
|
|
9 |
):
|
10 |
-
info = ModelInfo(simple_name, link, description)
|
11 |
-
|
12 |
for full_name in full_names:
|
13 |
model_info[full_name] = info
|
|
|
|
|
14 |
|
15 |
def get_model_info(name: str) -> ModelInfo:
|
16 |
if name in model_info:
|
@@ -18,7 +44,8 @@ def get_model_info(name: str) -> ModelInfo:
|
|
18 |
else:
|
19 |
# To fix this, please use `register_model_info` to register your model
|
20 |
return ModelInfo(
|
21 |
-
name, "", "Register the description at fastchat/model/model_registry.py"
|
|
|
22 |
)
|
23 |
|
24 |
def get_model_description_md(model_list):
|
@@ -50,6 +77,9 @@ register_model_info(
|
|
50 |
"LCM",
|
51 |
"https://huggingface.co/SimianLuo/LCM_Dreamshaper_v7",
|
52 |
"Latent Consistency Models.",
|
|
|
|
|
|
|
53 |
)
|
54 |
|
55 |
register_model_info(
|
@@ -57,27 +87,39 @@ register_model_info(
|
|
57 |
"LCM(v1.5/XL)",
|
58 |
"https://fal.ai/models/fast-lcm-diffusion-turbo",
|
59 |
"Latent Consistency Models (v1.5/XL)",
|
|
|
|
|
|
|
60 |
)
|
61 |
|
62 |
register_model_info(
|
63 |
["imagenhub_PlayGroundV2_generation", 'playground_PlayGroundV2_generation'],
|
64 |
-
"
|
65 |
"https://huggingface.co/playgroundai/playground-v2-1024px-aesthetic",
|
66 |
"Playground v2 – 1024px Aesthetic Model",
|
|
|
|
|
|
|
67 |
)
|
68 |
|
69 |
register_model_info(
|
70 |
["imagenhub_PlayGroundV2.5_generation", 'playground_PlayGroundV2.5_generation'],
|
71 |
-
"
|
72 |
"https://huggingface.co/playgroundai/playground-v2.5-1024px-aesthetic",
|
73 |
"Playground v2.5 is the state-of-the-art open-source model in aesthetic quality",
|
|
|
|
|
|
|
74 |
)
|
75 |
|
76 |
register_model_info(
|
77 |
["imagenhub_OpenJourney_generation"],
|
78 |
-
"
|
79 |
"https://huggingface.co/prompthero/openjourney",
|
80 |
"Openjourney is an open source Stable Diffusion fine tuned model on Midjourney images, by PromptHero.",
|
|
|
|
|
|
|
81 |
)
|
82 |
|
83 |
register_model_info(
|
@@ -85,6 +127,19 @@ register_model_info(
|
|
85 |
"SDXLTurbo",
|
86 |
"https://huggingface.co/stabilityai/sdxl-turbo",
|
87 |
"SDXL-Turbo is a fast generative text-to-image model.",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
88 |
)
|
89 |
|
90 |
register_model_info(
|
@@ -92,6 +147,9 @@ register_model_info(
|
|
92 |
"SDXL",
|
93 |
"https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0",
|
94 |
"SDXL is a Latent Diffusion Model that uses two fixed, pretrained text encoders.",
|
|
|
|
|
|
|
95 |
)
|
96 |
|
97 |
register_model_info(
|
@@ -99,6 +157,9 @@ register_model_info(
|
|
99 |
"SD3",
|
100 |
"https://huggingface.co/blog/sd3",
|
101 |
"SD3 is a novel Multimodal Diffusion Transformer (MMDiT) model.",
|
|
|
|
|
|
|
102 |
)
|
103 |
|
104 |
register_model_info(
|
@@ -106,6 +167,9 @@ register_model_info(
|
|
106 |
"PixArtAlpha",
|
107 |
"https://huggingface.co/PixArt-alpha/PixArt-XL-2-1024-MS",
|
108 |
"Pixart-α consists of pure transformer blocks for latent diffusion.",
|
|
|
|
|
|
|
109 |
)
|
110 |
|
111 |
register_model_info(
|
@@ -113,6 +177,9 @@ register_model_info(
|
|
113 |
"PixArtSigma",
|
114 |
"https://github.com/PixArt-alpha/PixArt-sigma",
|
115 |
"Improved version of Pixart-α.",
|
|
|
|
|
|
|
116 |
)
|
117 |
|
118 |
register_model_info(
|
@@ -120,6 +187,9 @@ register_model_info(
|
|
120 |
"SDXL-Lightning",
|
121 |
"https://huggingface.co/ByteDance/SDXL-Lightning",
|
122 |
"SDXL-Lightning is a lightning-fast text-to-image generation model.",
|
|
|
|
|
|
|
123 |
)
|
124 |
|
125 |
register_model_info(
|
@@ -127,6 +197,9 @@ register_model_info(
|
|
127 |
"StableCascade",
|
128 |
"https://huggingface.co/stabilityai/stable-cascade",
|
129 |
"StableCascade is built upon the Würstchen architecture and working at a much smaller latent space.",
|
|
|
|
|
|
|
130 |
)
|
131 |
|
132 |
register_model_info(
|
@@ -134,6 +207,9 @@ register_model_info(
|
|
134 |
"HunyuanDiT",
|
135 |
"https://github.com/Tencent/HunyuanDiT",
|
136 |
"HunyuanDiT is a Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding",
|
|
|
|
|
|
|
137 |
)
|
138 |
|
139 |
register_model_info(
|
@@ -141,6 +217,9 @@ register_model_info(
|
|
141 |
"Kolors",
|
142 |
"https://huggingface.co/Kwai-Kolors/Kolors",
|
143 |
"Kolors is a large-scale text-to-image generation model based on latent diffusion",
|
|
|
|
|
|
|
144 |
)
|
145 |
|
146 |
register_model_info(
|
@@ -148,6 +227,9 @@ register_model_info(
|
|
148 |
"AuraFlow",
|
149 |
"https://huggingface.co/fal/AuraFlow",
|
150 |
"Opensourced flow-based text-to-image generation model.",
|
|
|
|
|
|
|
151 |
)
|
152 |
|
153 |
register_model_info(
|
@@ -155,6 +237,9 @@ register_model_info(
|
|
155 |
"FLUX.1-schnell",
|
156 |
"https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux",
|
157 |
"Flux is a series of text-to-image generation models based on diffusion transformers. Timestep-distilled version.",
|
|
|
|
|
|
|
158 |
)
|
159 |
|
160 |
register_model_info(
|
@@ -162,6 +247,9 @@ register_model_info(
|
|
162 |
"FLUX.1-dev",
|
163 |
"https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux",
|
164 |
"Flux is a series of text-to-image generation models based on diffusion transformers. Guidance-distilled version.",
|
|
|
|
|
|
|
165 |
)
|
166 |
|
167 |
|
@@ -171,6 +259,9 @@ register_model_info(
|
|
171 |
"CycleDiffusion",
|
172 |
"https://github.com/ChenWu98/cycle-diffusion?tab=readme-ov-file",
|
173 |
"A latent space for stochastic diffusion models.",
|
|
|
|
|
|
|
174 |
)
|
175 |
|
176 |
register_model_info(
|
@@ -178,6 +269,9 @@ register_model_info(
|
|
178 |
"Pix2PixZero",
|
179 |
"https://pix2pixzero.github.io/",
|
180 |
"A zero-shot Image-to-Image translation model.",
|
|
|
|
|
|
|
181 |
)
|
182 |
|
183 |
register_model_info(
|
@@ -185,6 +279,9 @@ register_model_info(
|
|
185 |
"Prompt2prompt",
|
186 |
"https://prompt-to-prompt.github.io/",
|
187 |
"Image Editing with Cross-Attention Control.",
|
|
|
|
|
|
|
188 |
)
|
189 |
|
190 |
|
@@ -193,6 +290,9 @@ register_model_info(
|
|
193 |
"InstructPix2Pix",
|
194 |
"https://www.timothybrooks.com/instruct-pix2pix",
|
195 |
"An instruction-based image editing model.",
|
|
|
|
|
|
|
196 |
)
|
197 |
|
198 |
register_model_info(
|
@@ -200,6 +300,9 @@ register_model_info(
|
|
200 |
"MagicBrush",
|
201 |
"https://osu-nlp-group.github.io/MagicBrush/",
|
202 |
"Manually Annotated Dataset for Instruction-Guided Image Editing.",
|
|
|
|
|
|
|
203 |
)
|
204 |
|
205 |
register_model_info(
|
@@ -207,6 +310,9 @@ register_model_info(
|
|
207 |
"PNP",
|
208 |
"https://github.com/MichalGeyer/plug-and-play",
|
209 |
"Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation.",
|
|
|
|
|
|
|
210 |
)
|
211 |
|
212 |
register_model_info(
|
@@ -214,6 +320,9 @@ register_model_info(
|
|
214 |
"InfEdit",
|
215 |
"https://sled-group.github.io/InfEdit/",
|
216 |
"Inversion-Free Image Editing with Natural Language.",
|
|
|
|
|
|
|
217 |
)
|
218 |
|
219 |
register_model_info(
|
@@ -221,6 +330,9 @@ register_model_info(
|
|
221 |
"CosXLEdit",
|
222 |
"https://huggingface.co/stabilityai/cosxl",
|
223 |
"An instruction-based image editing model from SDXL.",
|
|
|
|
|
|
|
224 |
)
|
225 |
|
226 |
register_model_info(
|
@@ -228,6 +340,9 @@ register_model_info(
|
|
228 |
"UltraEdit",
|
229 |
"https://ultra-editing.github.io/",
|
230 |
"Instruction-based Fine-Grained Image Editing at Scale.",
|
|
|
|
|
|
|
231 |
)
|
232 |
|
233 |
register_model_info(
|
@@ -235,6 +350,9 @@ register_model_info(
|
|
235 |
"StableCascade",
|
236 |
"https://fal.ai/models/stable-cascade/api",
|
237 |
"StableCascade is a generative model that can generate high-quality images from text prompts.",
|
|
|
|
|
|
|
238 |
)
|
239 |
|
240 |
register_model_info(
|
@@ -242,6 +360,9 @@ register_model_info(
|
|
242 |
"AnimateDiff",
|
243 |
"https://fal.ai/models/fast-animatediff-t2v",
|
244 |
"AnimateDiff is a text-driven models that produce diverse and personalized animated images.",
|
|
|
|
|
|
|
245 |
)
|
246 |
|
247 |
register_model_info(
|
@@ -249,6 +370,9 @@ register_model_info(
|
|
249 |
"StableVideoDiffusion",
|
250 |
"https://fal.ai/models/fal-ai/fast-svd/text-to-video/api",
|
251 |
"Stable Video Diffusion empowers individuals to transform text and image inputs into vivid scenes.",
|
|
|
|
|
|
|
252 |
)
|
253 |
|
254 |
register_model_info(
|
@@ -256,6 +380,9 @@ register_model_info(
|
|
256 |
"AnimateDiff Turbo",
|
257 |
"https://fal.ai/models/fast-animatediff-t2v-turbo",
|
258 |
"AnimateDiff Turbo is a lightning version of AnimateDiff.",
|
|
|
|
|
|
|
259 |
)
|
260 |
|
261 |
register_model_info(
|
@@ -263,20 +390,28 @@ register_model_info(
|
|
263 |
"VideoCrafter2",
|
264 |
"https://ailab-cvc.github.io/videocrafter2/",
|
265 |
"VideoCrafter2 is a T2V model that disentangling motion from appearance.",
|
|
|
|
|
|
|
266 |
)
|
267 |
|
268 |
-
"""
|
269 |
register_model_info(
|
270 |
["videogenhub_LaVie_generation"],
|
271 |
"LaVie",
|
272 |
"https://github.com/Vchitect/LaVie",
|
273 |
"LaVie is a video generation model with cascaded latent diffusion models.",
|
|
|
|
|
|
|
274 |
)
|
275 |
register_model_info(
|
276 |
["videogenhub_ModelScope_generation"],
|
277 |
"ModelScope",
|
278 |
"https://arxiv.org/abs/2308.06571",
|
279 |
"ModelScope is a a T2V synthesis model that evolves from a T2I synthesis model.",
|
|
|
|
|
|
|
280 |
)
|
281 |
|
282 |
register_model_info(
|
@@ -284,13 +419,19 @@ register_model_info(
|
|
284 |
"OpenSora",
|
285 |
"https://github.com/hpcaitech/Open-Sora",
|
286 |
"A community-driven opensource implementation of Sora.",
|
|
|
|
|
|
|
287 |
)
|
288 |
-
|
289 |
register_model_info(
|
290 |
["videogenhub_OpenSora12_generation"],
|
291 |
"OpenSora v1.2",
|
292 |
"https://github.com/hpcaitech/Open-Sora",
|
293 |
"A community-driven opensource implementation of Sora. v1.2",
|
|
|
|
|
|
|
294 |
)
|
295 |
|
296 |
register_model_info(
|
@@ -298,6 +439,9 @@ register_model_info(
|
|
298 |
"CogVideoX",
|
299 |
"https://github.com/THUDM/CogVideo",
|
300 |
"Text-to-Video Diffusion Models with An Expert Transformer.",
|
|
|
|
|
|
|
301 |
)
|
302 |
|
303 |
register_model_info(
|
@@ -305,5 +449,10 @@ register_model_info(
|
|
305 |
"T2V-Turbo",
|
306 |
"https://github.com/Ji4chenLi/t2v-turbo",
|
307 |
"Video Consistency Model with Mixed Reward Feedback.",
|
|
|
|
|
|
|
308 |
)
|
309 |
|
|
|
|
|
|
1 |
from collections import namedtuple
|
2 |
from typing import List
|
3 |
|
4 |
+
IMAGE_GENERATION_MODELS = ['imagenhub_SDXLTurbo_generation','imagenhub_SDXL_generation', 'imagenhub_PixArtAlpha_generation', 'imagenhub_PixArtSigma_generation',
|
5 |
+
'imagenhub_OpenJourney_generation','imagenhub_SDXLLightning_generation', 'imagenhub_StableCascade_generation', 'imagenhub_HunyuanDiT_generation',
|
6 |
+
'playground_PlayGroundV2.5_generation', 'imagenhub_Kolors_generation', 'imagenhub_SD3_generation',
|
7 |
+
'fal_AuraFlow_text2image', 'fal_FLUX1schnell_text2image', 'fal_FLUX1dev_text2image'] # 'playground_PlayGroundV2_generation'
|
8 |
+
IMAGE_EDITION_MODELS = ['imagenhub_CycleDiffusion_edition', 'imagenhub_Pix2PixZero_edition', 'imagenhub_Prompt2prompt_edition',
|
9 |
+
'imagenhub_SDEdit_edition', 'imagenhub_InstructPix2Pix_edition',
|
10 |
+
'imagenhub_MagicBrush_edition', 'imagenhub_PNP_edition',
|
11 |
+
'imagenhub_InfEdit_edition', 'imagenhub_CosXLEdit_edition', 'imagenhub_UltraEdit_edition']
|
12 |
+
VIDEO_GENERATION_MODELS = ['fal_AnimateDiff_text2video',
|
13 |
+
'fal_AnimateDiffTurbo_text2video',
|
14 |
+
#'videogenhub_LaVie_generation',
|
15 |
+
'videogenhub_VideoCrafter2_generation',
|
16 |
+
#'videogenhub_ModelScope_generation',
|
17 |
+
'videogenhub_CogVideoX_generation', 'videogenhub_OpenSora12_generation',
|
18 |
+
#'videogenhub_OpenSora_generation',
|
19 |
+
#'videogenhub_T2VTurbo_generation',
|
20 |
+
'fal_T2VTurbo_text2video',
|
21 |
+
'fal_StableVideoDiffusion_text2video']
|
22 |
+
MUSEUM_UNSUPPORTED_MODELS = ['videogenhub_OpenSoraPlan_generation']
|
23 |
+
DESIRED_APPEAR_MODEL = ['videogenhub_T2VTurbo_generation','fal_StableVideoDiffusion_text2video']
|
24 |
+
|
25 |
+
ALL_MODELS = IMAGE_GENERATION_MODELS + IMAGE_EDITION_MODELS + VIDEO_GENERATION_MODELS
|
26 |
+
|
27 |
+
|
28 |
+
ModelInfo = namedtuple("ModelInfo", ["simple_name", "link", "description", "license", "organization", "type"])
|
29 |
model_info = {}
|
30 |
|
31 |
def register_model_info(
|
32 |
+
full_names: List[str], simple_name: str, link: str, description: str,
|
33 |
+
license: str, organization: str, model_type: str
|
34 |
):
|
35 |
+
info = ModelInfo(simple_name, link, description, license, organization, model_type)
|
|
|
36 |
for full_name in full_names:
|
37 |
model_info[full_name] = info
|
38 |
+
model_info[full_name.split("_")[1]] = info
|
39 |
+
model_info[simple_name] = info
|
40 |
|
41 |
def get_model_info(name: str) -> ModelInfo:
|
42 |
if name in model_info:
|
|
|
44 |
else:
|
45 |
# To fix this, please use `register_model_info` to register your model
|
46 |
return ModelInfo(
|
47 |
+
name, "-", "Register the description at fastchat/model/model_registry.py",
|
48 |
+
"-", "-", None
|
49 |
)
|
50 |
|
51 |
def get_model_description_md(model_list):
|
|
|
77 |
"LCM",
|
78 |
"https://huggingface.co/SimianLuo/LCM_Dreamshaper_v7",
|
79 |
"Latent Consistency Models.",
|
80 |
+
"MIT License",
|
81 |
+
"Tsinghua University",
|
82 |
+
"text2image_generation"
|
83 |
)
|
84 |
|
85 |
register_model_info(
|
|
|
87 |
"LCM(v1.5/XL)",
|
88 |
"https://fal.ai/models/fast-lcm-diffusion-turbo",
|
89 |
"Latent Consistency Models (v1.5/XL)",
|
90 |
+
"openrail++",
|
91 |
+
"Latent Consistency",
|
92 |
+
"text2image_generation"
|
93 |
)
|
94 |
|
95 |
register_model_info(
|
96 |
["imagenhub_PlayGroundV2_generation", 'playground_PlayGroundV2_generation'],
|
97 |
+
"PlayGround V2",
|
98 |
"https://huggingface.co/playgroundai/playground-v2-1024px-aesthetic",
|
99 |
"Playground v2 – 1024px Aesthetic Model",
|
100 |
+
"Playground v2 Community License",
|
101 |
+
"Playground",
|
102 |
+
"text2image_generation"
|
103 |
)
|
104 |
|
105 |
register_model_info(
|
106 |
["imagenhub_PlayGroundV2.5_generation", 'playground_PlayGroundV2.5_generation'],
|
107 |
+
"PlayGround V2.5",
|
108 |
"https://huggingface.co/playgroundai/playground-v2.5-1024px-aesthetic",
|
109 |
"Playground v2.5 is the state-of-the-art open-source model in aesthetic quality",
|
110 |
+
"Playground v2.5 Community License",
|
111 |
+
"Playground",
|
112 |
+
"text2image_generation"
|
113 |
)
|
114 |
|
115 |
register_model_info(
|
116 |
["imagenhub_OpenJourney_generation"],
|
117 |
+
"OpenJourney",
|
118 |
"https://huggingface.co/prompthero/openjourney",
|
119 |
"Openjourney is an open source Stable Diffusion fine tuned model on Midjourney images, by PromptHero.",
|
120 |
+
"creativeml-openrail-m",
|
121 |
+
"PromptHero",
|
122 |
+
"text2image_generation"
|
123 |
)
|
124 |
|
125 |
register_model_info(
|
|
|
127 |
"SDXLTurbo",
|
128 |
"https://huggingface.co/stabilityai/sdxl-turbo",
|
129 |
"SDXL-Turbo is a fast generative text-to-image model.",
|
130 |
+
"sai-nc-community (other)",
|
131 |
+
"Stability AI",
|
132 |
+
"text2image_generation"
|
133 |
+
)
|
134 |
+
|
135 |
+
register_model_info(
|
136 |
+
["imagenhub_SDEdit_edition"],
|
137 |
+
"SDEdit",
|
138 |
+
"https://sde-image-editing.github.io",
|
139 |
+
"SDEdit is an image synthesis and editing framework based on stochastic differential equations (SDEs) or diffusion models.",
|
140 |
+
"MIT License",
|
141 |
+
"Stanford University",
|
142 |
+
"image_edition"
|
143 |
)
|
144 |
|
145 |
register_model_info(
|
|
|
147 |
"SDXL",
|
148 |
"https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0",
|
149 |
"SDXL is a Latent Diffusion Model that uses two fixed, pretrained text encoders.",
|
150 |
+
"openrail++",
|
151 |
+
"Stability AI",
|
152 |
+
"text2image_generation"
|
153 |
)
|
154 |
|
155 |
register_model_info(
|
|
|
157 |
"SD3",
|
158 |
"https://huggingface.co/blog/sd3",
|
159 |
"SD3 is a novel Multimodal Diffusion Transformer (MMDiT) model.",
|
160 |
+
"stabilityai-nc-research-community",
|
161 |
+
"Stability AI",
|
162 |
+
"text2image_generation"
|
163 |
)
|
164 |
|
165 |
register_model_info(
|
|
|
167 |
"PixArtAlpha",
|
168 |
"https://huggingface.co/PixArt-alpha/PixArt-XL-2-1024-MS",
|
169 |
"Pixart-α consists of pure transformer blocks for latent diffusion.",
|
170 |
+
"openrail++",
|
171 |
+
"PixArt-alpha",
|
172 |
+
"text2image_generation"
|
173 |
)
|
174 |
|
175 |
register_model_info(
|
|
|
177 |
"PixArtSigma",
|
178 |
"https://github.com/PixArt-alpha/PixArt-sigma",
|
179 |
"Improved version of Pixart-α.",
|
180 |
+
"openrail++",
|
181 |
+
"PixArt-alpha",
|
182 |
+
"text2image_generation"
|
183 |
)
|
184 |
|
185 |
register_model_info(
|
|
|
187 |
"SDXL-Lightning",
|
188 |
"https://huggingface.co/ByteDance/SDXL-Lightning",
|
189 |
"SDXL-Lightning is a lightning-fast text-to-image generation model.",
|
190 |
+
"openrail++",
|
191 |
+
"ByteDance",
|
192 |
+
"text2image_generation"
|
193 |
)
|
194 |
|
195 |
register_model_info(
|
|
|
197 |
"StableCascade",
|
198 |
"https://huggingface.co/stabilityai/stable-cascade",
|
199 |
"StableCascade is built upon the Würstchen architecture and working at a much smaller latent space.",
|
200 |
+
"stable-cascade-nc-community (other)",
|
201 |
+
"Stability AI",
|
202 |
+
"text2image_generation"
|
203 |
)
|
204 |
|
205 |
register_model_info(
|
|
|
207 |
"HunyuanDiT",
|
208 |
"https://github.com/Tencent/HunyuanDiT",
|
209 |
"HunyuanDiT is a Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding",
|
210 |
+
"tencent-hunyuan-community",
|
211 |
+
"Tencent",
|
212 |
+
"text2image_generation"
|
213 |
)
|
214 |
|
215 |
register_model_info(
|
|
|
217 |
"Kolors",
|
218 |
"https://huggingface.co/Kwai-Kolors/Kolors",
|
219 |
"Kolors is a large-scale text-to-image generation model based on latent diffusion",
|
220 |
+
"Apache-2.0",
|
221 |
+
"Kwai Kolors",
|
222 |
+
"text2image_generation"
|
223 |
)
|
224 |
|
225 |
register_model_info(
|
|
|
227 |
"AuraFlow",
|
228 |
"https://huggingface.co/fal/AuraFlow",
|
229 |
"Opensourced flow-based text-to-image generation model.",
|
230 |
+
"Apache-2.0",
|
231 |
+
"Fal.AI",
|
232 |
+
"text2image_generation"
|
233 |
)
|
234 |
|
235 |
register_model_info(
|
|
|
237 |
"FLUX.1-schnell",
|
238 |
"https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux",
|
239 |
"Flux is a series of text-to-image generation models based on diffusion transformers. Timestep-distilled version.",
|
240 |
+
"flux-1-dev-non-commercial-license (other)",
|
241 |
+
"Black Forest Labs",
|
242 |
+
"text2image_generation"
|
243 |
)
|
244 |
|
245 |
register_model_info(
|
|
|
247 |
"FLUX.1-dev",
|
248 |
"https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux",
|
249 |
"Flux is a series of text-to-image generation models based on diffusion transformers. Guidance-distilled version.",
|
250 |
+
"flux-1-dev-non-commercial-license (other)",
|
251 |
+
"Black Forest Labs",
|
252 |
+
"text2image_generation"
|
253 |
)
|
254 |
|
255 |
|
|
|
259 |
"CycleDiffusion",
|
260 |
"https://github.com/ChenWu98/cycle-diffusion?tab=readme-ov-file",
|
261 |
"A latent space for stochastic diffusion models.",
|
262 |
+
"X11",
|
263 |
+
"Carnegie Mellon University",
|
264 |
+
"image_edition"
|
265 |
)
|
266 |
|
267 |
register_model_info(
|
|
|
269 |
"Pix2PixZero",
|
270 |
"https://pix2pixzero.github.io/",
|
271 |
"A zero-shot Image-to-Image translation model.",
|
272 |
+
"MIT License",
|
273 |
+
"Carnegie Mellon University, Adobe Research",
|
274 |
+
"image_edition"
|
275 |
)
|
276 |
|
277 |
register_model_info(
|
|
|
279 |
"Prompt2prompt",
|
280 |
"https://prompt-to-prompt.github.io/",
|
281 |
"Image Editing with Cross-Attention Control.",
|
282 |
+
"Apache-2.0",
|
283 |
+
"Google, Tel Aviv University",
|
284 |
+
"image_edition"
|
285 |
)
|
286 |
|
287 |
|
|
|
290 |
"InstructPix2Pix",
|
291 |
"https://www.timothybrooks.com/instruct-pix2pix",
|
292 |
"An instruction-based image editing model.",
|
293 |
+
"Copyright 2023 Timothy Brooks, Aleksander Holynski, Alexei A. Efros",
|
294 |
+
"University of California, Berkeley",
|
295 |
+
"image_edition"
|
296 |
)
|
297 |
|
298 |
register_model_info(
|
|
|
300 |
"MagicBrush",
|
301 |
"https://osu-nlp-group.github.io/MagicBrush/",
|
302 |
"Manually Annotated Dataset for Instruction-Guided Image Editing.",
|
303 |
+
"CC-BY-4.0",
|
304 |
+
"The Ohio State University, University of Waterloo",
|
305 |
+
"image_edition"
|
306 |
)
|
307 |
|
308 |
register_model_info(
|
|
|
310 |
"PNP",
|
311 |
"https://github.com/MichalGeyer/plug-and-play",
|
312 |
"Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation.",
|
313 |
+
"-",
|
314 |
+
"Weizmann Institute of Science",
|
315 |
+
"image_edition"
|
316 |
)
|
317 |
|
318 |
register_model_info(
|
|
|
320 |
"InfEdit",
|
321 |
"https://sled-group.github.io/InfEdit/",
|
322 |
"Inversion-Free Image Editing with Natural Language.",
|
323 |
+
"CC BY-NC-ND 4.0",
|
324 |
+
"University of Michigan, University of California, Berkeley",
|
325 |
+
"image_edition"
|
326 |
)
|
327 |
|
328 |
register_model_info(
|
|
|
330 |
"CosXLEdit",
|
331 |
"https://huggingface.co/stabilityai/cosxl",
|
332 |
"An instruction-based image editing model from SDXL.",
|
333 |
+
"cosxl-nc-community",
|
334 |
+
"Stability AI",
|
335 |
+
"image_edition"
|
336 |
)
|
337 |
|
338 |
register_model_info(
|
|
|
340 |
"UltraEdit",
|
341 |
"https://ultra-editing.github.io/",
|
342 |
"Instruction-based Fine-Grained Image Editing at Scale.",
|
343 |
+
"other",
|
344 |
+
"Peking University; BIGAI",
|
345 |
+
"image_edition"
|
346 |
)
|
347 |
|
348 |
register_model_info(
|
|
|
350 |
"StableCascade",
|
351 |
"https://fal.ai/models/stable-cascade/api",
|
352 |
"StableCascade is a generative model that can generate high-quality images from text prompts.",
|
353 |
+
"stable-cascade-nc-community (other)",
|
354 |
+
"Stability AI",
|
355 |
+
"image_edition"
|
356 |
)
|
357 |
|
358 |
register_model_info(
|
|
|
360 |
"AnimateDiff",
|
361 |
"https://fal.ai/models/fast-animatediff-t2v",
|
362 |
"AnimateDiff is a text-driven models that produce diverse and personalized animated images.",
|
363 |
+
"creativeml-openrail-m",
|
364 |
+
"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",
|
365 |
+
"text2video_generation"
|
366 |
)
|
367 |
|
368 |
register_model_info(
|
|
|
370 |
"StableVideoDiffusion",
|
371 |
"https://fal.ai/models/fal-ai/fast-svd/text-to-video/api",
|
372 |
"Stable Video Diffusion empowers individuals to transform text and image inputs into vivid scenes.",
|
373 |
+
"SVD-nc-community",
|
374 |
+
"Stability AI",
|
375 |
+
"text2video_generation"
|
376 |
)
|
377 |
|
378 |
register_model_info(
|
|
|
380 |
"AnimateDiff Turbo",
|
381 |
"https://fal.ai/models/fast-animatediff-t2v-turbo",
|
382 |
"AnimateDiff Turbo is a lightning version of AnimateDiff.",
|
383 |
+
"creativeml-openrail-m",
|
384 |
+
"The Chinese University of Hong Kong, Shanghai AI Lab, Stanford University",
|
385 |
+
"text2video_generation"
|
386 |
)
|
387 |
|
388 |
register_model_info(
|
|
|
390 |
"VideoCrafter2",
|
391 |
"https://ailab-cvc.github.io/videocrafter2/",
|
392 |
"VideoCrafter2 is a T2V model that disentangling motion from appearance.",
|
393 |
+
"Apache 2.0",
|
394 |
+
"Tencent AI Lab",
|
395 |
+
"text2video_generation"
|
396 |
)
|
397 |
|
|
|
398 |
register_model_info(
|
399 |
["videogenhub_LaVie_generation"],
|
400 |
"LaVie",
|
401 |
"https://github.com/Vchitect/LaVie",
|
402 |
"LaVie is a video generation model with cascaded latent diffusion models.",
|
403 |
+
"Apache 2.0",
|
404 |
+
"Shanghai AI Lab",
|
405 |
+
"text2video_generation"
|
406 |
)
|
407 |
register_model_info(
|
408 |
["videogenhub_ModelScope_generation"],
|
409 |
"ModelScope",
|
410 |
"https://arxiv.org/abs/2308.06571",
|
411 |
"ModelScope is a a T2V synthesis model that evolves from a T2I synthesis model.",
|
412 |
+
"cc-by-nc-4.0",
|
413 |
+
"Alibaba Group",
|
414 |
+
"text2video_generation"
|
415 |
)
|
416 |
|
417 |
register_model_info(
|
|
|
419 |
"OpenSora",
|
420 |
"https://github.com/hpcaitech/Open-Sora",
|
421 |
"A community-driven opensource implementation of Sora.",
|
422 |
+
"Apache 2.0",
|
423 |
+
"HPC-AI Tech",
|
424 |
+
"text2video_generation"
|
425 |
)
|
426 |
+
|
427 |
register_model_info(
|
428 |
["videogenhub_OpenSora12_generation"],
|
429 |
"OpenSora v1.2",
|
430 |
"https://github.com/hpcaitech/Open-Sora",
|
431 |
"A community-driven opensource implementation of Sora. v1.2",
|
432 |
+
"Apache 2.0",
|
433 |
+
"HPC-AI Tech",
|
434 |
+
"text2video_generation"
|
435 |
)
|
436 |
|
437 |
register_model_info(
|
|
|
439 |
"CogVideoX",
|
440 |
"https://github.com/THUDM/CogVideo",
|
441 |
"Text-to-Video Diffusion Models with An Expert Transformer.",
|
442 |
+
"CogVideoX LICENSE",
|
443 |
+
"THUDM",
|
444 |
+
"text2video_generation"
|
445 |
)
|
446 |
|
447 |
register_model_info(
|
|
|
449 |
"T2V-Turbo",
|
450 |
"https://github.com/Ji4chenLi/t2v-turbo",
|
451 |
"Video Consistency Model with Mixed Reward Feedback.",
|
452 |
+
"cc-by-nc-4.0",
|
453 |
+
"University of California, Santa Barbara",
|
454 |
+
"text2video_generation"
|
455 |
)
|
456 |
|
457 |
+
|
458 |
+
assert all([model in model_info for model in ALL_MODELS]), "Some models are not registered in model_info"
|
model/models/__init__.py
CHANGED
@@ -2,6 +2,7 @@ from .imagenhub_models import load_imagenhub_model
|
|
2 |
from .playground_api import load_playground_model
|
3 |
from .fal_api_models import load_fal_model
|
4 |
from .videogenhub_models import load_videogenhub_model
|
|
|
5 |
|
6 |
|
7 |
# IMAGE_GENERATION_MODELS = ['fal_LCM(v1.5/XL)_text2image','fal_SDXLTurbo_text2image','fal_SDXL_text2image', 'imagenhub_PixArtAlpha_generation', 'fal_PixArtSigma_text2image',
|
@@ -28,6 +29,12 @@ VIDEO_GENERATION_MODELS = ['fal_AnimateDiff_text2video',
|
|
28 |
MUSEUM_UNSUPPORTED_MODELS = ['videogenhub_OpenSoraPlan_generation']
|
29 |
DESIRED_APPEAR_MODEL = ['videogenhub_T2VTurbo_generation','fal_StableVideoDiffusion_text2video']
|
30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
def load_pipeline(model_name):
|
32 |
"""
|
33 |
Load a model pipeline based on the model name
|
|
|
2 |
from .playground_api import load_playground_model
|
3 |
from .fal_api_models import load_fal_model
|
4 |
from .videogenhub_models import load_videogenhub_model
|
5 |
+
from ..model_registry import model_info
|
6 |
|
7 |
|
8 |
# IMAGE_GENERATION_MODELS = ['fal_LCM(v1.5/XL)_text2image','fal_SDXLTurbo_text2image','fal_SDXL_text2image', 'imagenhub_PixArtAlpha_generation', 'fal_PixArtSigma_text2image',
|
|
|
29 |
MUSEUM_UNSUPPORTED_MODELS = ['videogenhub_OpenSoraPlan_generation']
|
30 |
DESIRED_APPEAR_MODEL = ['videogenhub_T2VTurbo_generation','fal_StableVideoDiffusion_text2video']
|
31 |
|
32 |
+
ALL_MODELS = IMAGE_GENERATION_MODELS + IMAGE_EDITION_MODELS + VIDEO_GENERATION_MODELS
|
33 |
+
|
34 |
+
missing_models = [model for model in ALL_MODELS if model not in model_info]
|
35 |
+
if missing_models:
|
36 |
+
raise ValueError(f"Missing models in model_info: {missing_models}")
|
37 |
+
|
38 |
def load_pipeline(model_name):
|
39 |
"""
|
40 |
Load a model pipeline based on the model name
|