Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -28,6 +28,23 @@ models_and_languages = {
|
|
28 |
"croissantllm/CroissantLLMBase": ["en", "fr"]
|
29 |
}
|
30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
# Variables globales
|
32 |
model = None
|
33 |
tokenizer = None
|
@@ -60,9 +77,18 @@ def load_model(model_name, progress=gr.Progress()):
|
|
60 |
|
61 |
progress(1.0, desc="Modèle chargé")
|
62 |
available_languages = models_and_languages[model_name]
|
63 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
except Exception as e:
|
65 |
-
return f"Erreur lors du chargement du modèle : {str(e)}", gr.Dropdown(visible=False)
|
66 |
|
67 |
def set_language(lang):
|
68 |
global selected_language
|
@@ -177,7 +203,7 @@ def reset():
|
|
177 |
return "", 1.0, 1.0, 50, None, None, None, None, gr.Dropdown(visible=False), ""
|
178 |
|
179 |
with gr.Blocks() as demo:
|
180 |
-
gr.Markdown("#
|
181 |
|
182 |
with gr.Accordion("Sélection du modèle"):
|
183 |
model_dropdown = gr.Dropdown(choices=list(models_and_languages.keys()), label="Choisissez un modèle")
|
@@ -205,7 +231,9 @@ with gr.Blocks() as demo:
|
|
205 |
|
206 |
reset_button = gr.Button("Réinitialiser")
|
207 |
|
208 |
-
load_button.click(load_model,
|
|
|
|
|
209 |
language_dropdown.change(set_language, inputs=[language_dropdown], outputs=[language_output])
|
210 |
analyze_button.click(analyze_next_token,
|
211 |
inputs=[input_text, temperature, top_p, top_k],
|
|
|
28 |
"croissantllm/CroissantLLMBase": ["en", "fr"]
|
29 |
}
|
30 |
|
31 |
+
# Paramètres recommandés pour chaque modèle
|
32 |
+
model_parameters = {
|
33 |
+
"meta-llama/Llama-2-13b-hf": {"temperature": 0.8, "top_p": 0.9, "top_k": 40},
|
34 |
+
"meta-llama/Llama-2-7b-hf": {"temperature": 0.8, "top_p": 0.9, "top_k": 40},
|
35 |
+
"meta-llama/Llama-2-70b-hf": {"temperature": 0.8, "top_p": 0.9, "top_k": 40},
|
36 |
+
"meta-llama/Meta-Llama-3-8B": {"temperature": 0.75, "top_p": 0.9, "top_k": 50},
|
37 |
+
"meta-llama/Llama-3.2-3B": {"temperature": 0.75, "top_p": 0.9, "top_k": 50},
|
38 |
+
"meta-llama/Llama-3.1-8B": {"temperature": 0.75, "top_p": 0.9, "top_k": 50},
|
39 |
+
"mistralai/Mistral-7B-v0.1": {"temperature": 0.7, "top_p": 0.9, "top_k": 50},
|
40 |
+
"mistralai/Mixtral-8x7B-v0.1": {"temperature": 0.8, "top_p": 0.95, "top_k": 50},
|
41 |
+
"mistralai/Mistral-7B-v0.3": {"temperature": 0.7, "top_p": 0.9, "top_k": 50},
|
42 |
+
"google/gemma-2-2b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
43 |
+
"google/gemma-2-9b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
44 |
+
"google/gemma-2-27b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
45 |
+
"croissantllm/CroissantLLMBase": {"temperature": 0.8, "top_p": 0.92, "top_k": 50}
|
46 |
+
}
|
47 |
+
|
48 |
# Variables globales
|
49 |
model = None
|
50 |
tokenizer = None
|
|
|
77 |
|
78 |
progress(1.0, desc="Modèle chargé")
|
79 |
available_languages = models_and_languages[model_name]
|
80 |
+
|
81 |
+
# Mise à jour des sliders avec les valeurs recommandées
|
82 |
+
params = model_parameters[model_name]
|
83 |
+
return (
|
84 |
+
f"Modèle {model_name} chargé avec succès. Langues disponibles : {', '.join(available_languages)}",
|
85 |
+
gr.Dropdown(choices=available_languages, value=available_languages[0], visible=True, interactive=True),
|
86 |
+
params["temperature"],
|
87 |
+
params["top_p"],
|
88 |
+
params["top_k"]
|
89 |
+
)
|
90 |
except Exception as e:
|
91 |
+
return f"Erreur lors du chargement du modèle : {str(e)}", gr.Dropdown(visible=False), None, None, None
|
92 |
|
93 |
def set_language(lang):
|
94 |
global selected_language
|
|
|
203 |
return "", 1.0, 1.0, 50, None, None, None, None, gr.Dropdown(visible=False), ""
|
204 |
|
205 |
with gr.Blocks() as demo:
|
206 |
+
gr.Markdown("# LLM&BIAS")
|
207 |
|
208 |
with gr.Accordion("Sélection du modèle"):
|
209 |
model_dropdown = gr.Dropdown(choices=list(models_and_languages.keys()), label="Choisissez un modèle")
|
|
|
231 |
|
232 |
reset_button = gr.Button("Réinitialiser")
|
233 |
|
234 |
+
load_button.click(load_model,
|
235 |
+
inputs=[model_dropdown],
|
236 |
+
outputs=[load_output, language_dropdown, temperature, top_p, top_k])
|
237 |
language_dropdown.change(set_language, inputs=[language_dropdown], outputs=[language_output])
|
238 |
analyze_button.click(analyze_next_token,
|
239 |
inputs=[input_text, temperature, top_p, top_k],
|