Spaces:
Sleeping
Sleeping
Commit
·
52fc709
1
Parent(s):
c4d55d5
Fix: when using http worker, only download if inference is on localhost
Browse files
utils.py
CHANGED
|
@@ -36,7 +36,7 @@ LLM_MODEL_PATH = env.get("LLM_MODEL_PATH", None)
|
|
| 36 |
MAX_TOKENS = int(env.get("MAX_TOKENS", 1000))
|
| 37 |
TEMPERATURE = float(env.get("TEMPERATURE", 0.3))
|
| 38 |
|
| 39 |
-
performing_local_inference = (LLM_WORKER == "in_memory" or LLM_WORKER == "http")
|
| 40 |
|
| 41 |
if LLM_MODEL_PATH and len(LLM_MODEL_PATH) > 0:
|
| 42 |
print(f"Using local model from {LLM_MODEL_PATH}")
|
|
|
|
| 36 |
MAX_TOKENS = int(env.get("MAX_TOKENS", 1000))
|
| 37 |
TEMPERATURE = float(env.get("TEMPERATURE", 0.3))
|
| 38 |
|
| 39 |
+
performing_local_inference = (LLM_WORKER == "in_memory" or (LLM_WORKER == "http" and "localhost" in URL))
|
| 40 |
|
| 41 |
if LLM_MODEL_PATH and len(LLM_MODEL_PATH) > 0:
|
| 42 |
print(f"Using local model from {LLM_MODEL_PATH}")
|