fixed bug on running openllm
Browse files- .env.example +1 -1
- Makefile +3 -0
- app_modules/llm_loader.py +1 -1
.env.example
CHANGED
|
@@ -7,7 +7,7 @@ LLM_MODEL_TYPE=huggingface
|
|
| 7 |
# LLM_MODEL_TYPE=stablelm
|
| 8 |
# LLM_MODEL_TYPE=openllm
|
| 9 |
|
| 10 |
-
OPENLLM_SERVER_URL=
|
| 11 |
|
| 12 |
OPENAI_API_KEY=
|
| 13 |
|
|
|
|
| 7 |
# LLM_MODEL_TYPE=stablelm
|
| 8 |
# LLM_MODEL_TYPE=openllm
|
| 9 |
|
| 10 |
+
OPENLLM_SERVER_URL=http://localhost:64300
|
| 11 |
|
| 12 |
OPENAI_API_KEY=
|
| 13 |
|
Makefile
CHANGED
|
@@ -25,6 +25,9 @@ else
|
|
| 25 |
openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf --port=${PORT}
|
| 26 |
endif
|
| 27 |
|
|
|
|
|
|
|
|
|
|
| 28 |
ingest:
|
| 29 |
python ingest.py
|
| 30 |
|
|
|
|
| 25 |
openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf --port=${PORT}
|
| 26 |
endif
|
| 27 |
|
| 28 |
+
openllm-cpu:
|
| 29 |
+
CUDA_VISIBLE_DEVICES="" openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf
|
| 30 |
+
|
| 31 |
ingest:
|
| 32 |
python ingest.py
|
| 33 |
|
app_modules/llm_loader.py
CHANGED
|
@@ -161,7 +161,7 @@ class LLMLoader:
|
|
| 161 |
print(f" server url: {server_url}")
|
| 162 |
self.llm = OpenLLM(
|
| 163 |
server_url=server_url,
|
| 164 |
-
callbacks=callbacks,
|
| 165 |
verbose=True,
|
| 166 |
)
|
| 167 |
elif self.llm_model_type.startswith("gpt4all"):
|
|
|
|
| 161 |
print(f" server url: {server_url}")
|
| 162 |
self.llm = OpenLLM(
|
| 163 |
server_url=server_url,
|
| 164 |
+
# callbacks=callbacks,
|
| 165 |
verbose=True,
|
| 166 |
)
|
| 167 |
elif self.llm_model_type.startswith("gpt4all"):
|