Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -28,6 +28,23 @@ models_and_languages = {
|
|
| 28 |
"croissantllm/CroissantLLMBase": ["en", "fr"]
|
| 29 |
}
|
| 30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
# Variables globales
|
| 32 |
model = None
|
| 33 |
tokenizer = None
|
|
@@ -60,9 +77,18 @@ def load_model(model_name, progress=gr.Progress()):
|
|
| 60 |
|
| 61 |
progress(1.0, desc="Modèle chargé")
|
| 62 |
available_languages = models_and_languages[model_name]
|
| 63 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 64 |
except Exception as e:
|
| 65 |
-
return f"Erreur lors du chargement du modèle : {str(e)}", gr.Dropdown(visible=False)
|
| 66 |
|
| 67 |
def set_language(lang):
|
| 68 |
global selected_language
|
|
@@ -177,7 +203,7 @@ def reset():
|
|
| 177 |
return "", 1.0, 1.0, 50, None, None, None, None, gr.Dropdown(visible=False), ""
|
| 178 |
|
| 179 |
with gr.Blocks() as demo:
|
| 180 |
-
gr.Markdown("#
|
| 181 |
|
| 182 |
with gr.Accordion("Sélection du modèle"):
|
| 183 |
model_dropdown = gr.Dropdown(choices=list(models_and_languages.keys()), label="Choisissez un modèle")
|
|
@@ -205,7 +231,9 @@ with gr.Blocks() as demo:
|
|
| 205 |
|
| 206 |
reset_button = gr.Button("Réinitialiser")
|
| 207 |
|
| 208 |
-
load_button.click(load_model,
|
|
|
|
|
|
|
| 209 |
language_dropdown.change(set_language, inputs=[language_dropdown], outputs=[language_output])
|
| 210 |
analyze_button.click(analyze_next_token,
|
| 211 |
inputs=[input_text, temperature, top_p, top_k],
|
|
|
|
| 28 |
"croissantllm/CroissantLLMBase": ["en", "fr"]
|
| 29 |
}
|
| 30 |
|
| 31 |
+
# Paramètres recommandés pour chaque modèle
|
| 32 |
+
model_parameters = {
|
| 33 |
+
"meta-llama/Llama-2-13b-hf": {"temperature": 0.8, "top_p": 0.9, "top_k": 40},
|
| 34 |
+
"meta-llama/Llama-2-7b-hf": {"temperature": 0.8, "top_p": 0.9, "top_k": 40},
|
| 35 |
+
"meta-llama/Llama-2-70b-hf": {"temperature": 0.8, "top_p": 0.9, "top_k": 40},
|
| 36 |
+
"meta-llama/Meta-Llama-3-8B": {"temperature": 0.75, "top_p": 0.9, "top_k": 50},
|
| 37 |
+
"meta-llama/Llama-3.2-3B": {"temperature": 0.75, "top_p": 0.9, "top_k": 50},
|
| 38 |
+
"meta-llama/Llama-3.1-8B": {"temperature": 0.75, "top_p": 0.9, "top_k": 50},
|
| 39 |
+
"mistralai/Mistral-7B-v0.1": {"temperature": 0.7, "top_p": 0.9, "top_k": 50},
|
| 40 |
+
"mistralai/Mixtral-8x7B-v0.1": {"temperature": 0.8, "top_p": 0.95, "top_k": 50},
|
| 41 |
+
"mistralai/Mistral-7B-v0.3": {"temperature": 0.7, "top_p": 0.9, "top_k": 50},
|
| 42 |
+
"google/gemma-2-2b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
| 43 |
+
"google/gemma-2-9b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
| 44 |
+
"google/gemma-2-27b": {"temperature": 0.7, "top_p": 0.95, "top_k": 40},
|
| 45 |
+
"croissantllm/CroissantLLMBase": {"temperature": 0.8, "top_p": 0.92, "top_k": 50}
|
| 46 |
+
}
|
| 47 |
+
|
| 48 |
# Variables globales
|
| 49 |
model = None
|
| 50 |
tokenizer = None
|
|
|
|
| 77 |
|
| 78 |
progress(1.0, desc="Modèle chargé")
|
| 79 |
available_languages = models_and_languages[model_name]
|
| 80 |
+
|
| 81 |
+
# Mise à jour des sliders avec les valeurs recommandées
|
| 82 |
+
params = model_parameters[model_name]
|
| 83 |
+
return (
|
| 84 |
+
f"Modèle {model_name} chargé avec succès. Langues disponibles : {', '.join(available_languages)}",
|
| 85 |
+
gr.Dropdown(choices=available_languages, value=available_languages[0], visible=True, interactive=True),
|
| 86 |
+
params["temperature"],
|
| 87 |
+
params["top_p"],
|
| 88 |
+
params["top_k"]
|
| 89 |
+
)
|
| 90 |
except Exception as e:
|
| 91 |
+
return f"Erreur lors du chargement du modèle : {str(e)}", gr.Dropdown(visible=False), None, None, None
|
| 92 |
|
| 93 |
def set_language(lang):
|
| 94 |
global selected_language
|
|
|
|
| 203 |
return "", 1.0, 1.0, 50, None, None, None, None, gr.Dropdown(visible=False), ""
|
| 204 |
|
| 205 |
with gr.Blocks() as demo:
|
| 206 |
+
gr.Markdown("# LLM&BIAS")
|
| 207 |
|
| 208 |
with gr.Accordion("Sélection du modèle"):
|
| 209 |
model_dropdown = gr.Dropdown(choices=list(models_and_languages.keys()), label="Choisissez un modèle")
|
|
|
|
| 231 |
|
| 232 |
reset_button = gr.Button("Réinitialiser")
|
| 233 |
|
| 234 |
+
load_button.click(load_model,
|
| 235 |
+
inputs=[model_dropdown],
|
| 236 |
+
outputs=[load_output, language_dropdown, temperature, top_p, top_k])
|
| 237 |
language_dropdown.change(set_language, inputs=[language_dropdown], outputs=[language_output])
|
| 238 |
analyze_button.click(analyze_next_token,
|
| 239 |
inputs=[input_text, temperature, top_p, top_k],
|