Update app.py
Browse files
app.py
CHANGED
|
@@ -42,7 +42,7 @@ def cache_answer(question_id: str, answer: str):
|
|
| 42 |
|
| 43 |
|
| 44 |
# --- Model Setup ---
|
| 45 |
-
MODEL_NAME = 'Qwen/Qwen2.5-3B-Instruct' # 'meta-llama/Llama-3.2-3B-Instruct'
|
| 46 |
|
| 47 |
|
| 48 |
# "Qwen/Qwen2.5-VL-3B-Instruct"#'meta-llama/Llama-2-7b-hf'#'meta-llama/Llama-3.1-8B-Instruct'#'TinyLlama/TinyLlama-1.1B-Chat-v1.0'#'mistralai/Mistral-7B-Instruct-v0.2'#'microsoft/DialoGPT-small'# 'EleutherAI/gpt-neo-2.7B'#'distilbert/distilgpt2'#'deepseek-ai/DeepSeek-R1-Distill-Qwen-7B'#'mistralai/Mistral-7B-Instruct-v0.2'
|
|
@@ -70,8 +70,8 @@ def load_model(model_name):
|
|
| 70 |
|
| 71 |
# Load the model and tokenizer locally
|
| 72 |
# model, tokenizer = load_model()
|
| 73 |
-
|
| 74 |
-
model_id = "
|
| 75 |
model = HfApiModel(model_id)
|
| 76 |
#from smolagents import TransformersModel
|
| 77 |
# model = TransformersModel(
|
|
|
|
| 42 |
|
| 43 |
|
| 44 |
# --- Model Setup ---
|
| 45 |
+
#MODEL_NAME = 'Qwen/Qwen2.5-3B-Instruct' # 'meta-llama/Llama-3.2-3B-Instruct'
|
| 46 |
|
| 47 |
|
| 48 |
# "Qwen/Qwen2.5-VL-3B-Instruct"#'meta-llama/Llama-2-7b-hf'#'meta-llama/Llama-3.1-8B-Instruct'#'TinyLlama/TinyLlama-1.1B-Chat-v1.0'#'mistralai/Mistral-7B-Instruct-v0.2'#'microsoft/DialoGPT-small'# 'EleutherAI/gpt-neo-2.7B'#'distilbert/distilgpt2'#'deepseek-ai/DeepSeek-R1-Distill-Qwen-7B'#'mistralai/Mistral-7B-Instruct-v0.2'
|
|
|
|
| 70 |
|
| 71 |
# Load the model and tokenizer locally
|
| 72 |
# model, tokenizer = load_model()
|
| 73 |
+
|
| 74 |
+
model_id = "meta-llama/Llama-3.1-8B-Instruct" # "microsoft/phi-2"# not working out of the box"google/gemma-2-2b-it" #toobig"Qwen/Qwen1.5-7B-Chat"#working but stupid: "meta-llama/Llama-3.2-3B-Instruct"
|
| 75 |
model = HfApiModel(model_id)
|
| 76 |
#from smolagents import TransformersModel
|
| 77 |
# model = TransformersModel(
|