Spaces:
Paused
Paused
NGUYEN, Xuan Phi
commited on
Commit
·
e9cbae4
1
Parent(s):
e33248e
update
Browse files
app.py
CHANGED
|
@@ -697,23 +697,23 @@ MODEL_PATH = os.environ.get("MODEL_PATH", "notfound, please set `export MODEL_PA
|
|
| 697 |
|
| 698 |
def launch():
|
| 699 |
global demo, llm, DEBUG
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 700 |
if DEBUG:
|
| 701 |
-
model_desc
|
| 702 |
response_fn = debug_chat_response_echo
|
| 703 |
else:
|
| 704 |
-
model_desc = MODEL_DESC
|
| 705 |
-
model_path = MODEL_PATH
|
| 706 |
-
assert os.path.exists(model_path), f'{model_path} not found'
|
| 707 |
-
model_title = MODEL_TITLE
|
| 708 |
-
tensor_parallel = TENSOR_PARALLEL
|
| 709 |
-
assert tensor_parallel > 0 , f'{tensor_parallel} invalid'
|
| 710 |
-
dtype = DTYPE
|
| 711 |
-
|
| 712 |
# ! load the model
|
| 713 |
llm = LLM(model=model_path, dtype=dtype, tensor_parallel_size=tensor_parallel)
|
| 714 |
|
| 715 |
-
sys_prompt = SYSTEM_PROMPT_1
|
| 716 |
-
max_tokens = 4096
|
| 717 |
print(f'Use system prompt:\n{sys_prompt}')
|
| 718 |
|
| 719 |
# response_fn = chat_response_stream_multiturn if args.multiturn else chat_response_stream
|
|
|
|
| 697 |
|
| 698 |
def launch():
|
| 699 |
global demo, llm, DEBUG
|
| 700 |
+
model_desc = MODEL_DESC
|
| 701 |
+
model_path = MODEL_PATH
|
| 702 |
+
assert os.path.exists(model_path), f'{model_path} not found'
|
| 703 |
+
model_title = MODEL_TITLE
|
| 704 |
+
tensor_parallel = TENSOR_PARALLEL
|
| 705 |
+
assert tensor_parallel > 0 , f'{tensor_parallel} invalid'
|
| 706 |
+
dtype = DTYPE
|
| 707 |
+
sys_prompt = SYSTEM_PROMPT_1
|
| 708 |
+
max_tokens = 4096
|
| 709 |
+
|
| 710 |
if DEBUG:
|
| 711 |
+
model_desc += "<br>!!!!! This is in debug mode, responses will be copy original"
|
| 712 |
response_fn = debug_chat_response_echo
|
| 713 |
else:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 714 |
# ! load the model
|
| 715 |
llm = LLM(model=model_path, dtype=dtype, tensor_parallel_size=tensor_parallel)
|
| 716 |
|
|
|
|
|
|
|
| 717 |
print(f'Use system prompt:\n{sys_prompt}')
|
| 718 |
|
| 719 |
# response_fn = chat_response_stream_multiturn if args.multiturn else chat_response_stream
|