Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -25,10 +25,9 @@ from io import StringIO
|
|
| 25 |
last = 0
|
| 26 |
CHROMA_DATA_PATH = "chroma_data/"
|
| 27 |
EMBED_MODEL = "sentence-transformers/all-MiniLM-L6-v2" # "BAAI/bge-m3"
|
| 28 |
-
LLM_NAME = "mistralai/Mistral-Nemo-Instruct-2407"
|
| 29 |
-
LLM_NAME = "swiss-ai/Apertus-8B-Instruct-2509"
|
| 30 |
#LLM_NAME = "W4D/YugoGPT-7B-Instruct-GGUF"
|
| 31 |
-
# all-MiniLM-L6-v2
|
| 32 |
CHUNK_SIZE = 800
|
| 33 |
CHUNK_OVERLAP = 50
|
| 34 |
max_results = 3
|
|
@@ -77,9 +76,10 @@ Settings.llm = HuggingFaceInferenceAPI(model_name=LLM_NAME,
|
|
| 77 |
# stopping_ids=[50278, 50279, 50277, 1, 0],
|
| 78 |
generate_kwargs={"temperature": temperature, "top_p":top_p, "repetition_penalty": repetition_penalty,
|
| 79 |
"presence_penalty": presence_penalty, "frequency_penalty": frequency_penalty,
|
| 80 |
-
"top_k": top_k, "do_sample": False
|
| 81 |
# tokenizer_kwargs={"max_length": 4096},
|
| 82 |
tokenizer_name=LLM_NAME,
|
|
|
|
| 83 |
# provider="auto",
|
| 84 |
# task="None"
|
| 85 |
)
|
|
|
|
| 25 |
last = 0
|
| 26 |
CHROMA_DATA_PATH = "chroma_data/"
|
| 27 |
EMBED_MODEL = "sentence-transformers/all-MiniLM-L6-v2" # "BAAI/bge-m3"
|
| 28 |
+
#LLM_NAME = "mistralai/Mistral-Nemo-Instruct-2407"
|
| 29 |
+
LLM_NAME = "swiss-ai/Apertus-8B-Instruct-2509:publicai"
|
| 30 |
#LLM_NAME = "W4D/YugoGPT-7B-Instruct-GGUF"
|
|
|
|
| 31 |
CHUNK_SIZE = 800
|
| 32 |
CHUNK_OVERLAP = 50
|
| 33 |
max_results = 3
|
|
|
|
| 76 |
# stopping_ids=[50278, 50279, 50277, 1, 0],
|
| 77 |
generate_kwargs={"temperature": temperature, "top_p":top_p, "repetition_penalty": repetition_penalty,
|
| 78 |
"presence_penalty": presence_penalty, "frequency_penalty": frequency_penalty,
|
| 79 |
+
"top_k": top_k, "do_sample": False },
|
| 80 |
# tokenizer_kwargs={"max_length": 4096},
|
| 81 |
tokenizer_name=LLM_NAME,
|
| 82 |
+
api_key="hf_" + HF_TOKEN,
|
| 83 |
# provider="auto",
|
| 84 |
# task="None"
|
| 85 |
)
|