Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -54,6 +54,7 @@ from huggingface_hub import HfApi
|
|
| 54 |
from huggingface_hub import login
|
| 55 |
from huggingface_hub import revision_exists
|
| 56 |
|
|
|
|
| 57 |
api = HfApi()
|
| 58 |
repo_name = "paralym/pangea_logs"
|
| 59 |
|
|
@@ -213,7 +214,7 @@ def add_message(history, message):
|
|
| 213 |
|
| 214 |
|
| 215 |
@spaces.GPU
|
| 216 |
-
def bot(history):
|
| 217 |
print("### turn start history",history)
|
| 218 |
print("### turn start conv",our_chatbot.conversation)
|
| 219 |
text = history[-1][0]
|
|
@@ -350,10 +351,10 @@ def bot(history):
|
|
| 350 |
inputs=input_ids,
|
| 351 |
streamer=streamer,
|
| 352 |
images=image_tensor,
|
| 353 |
-
max_new_tokens=1024,
|
| 354 |
do_sample=True,
|
| 355 |
-
temperature=
|
| 356 |
-
|
|
|
|
| 357 |
use_cache=False,
|
| 358 |
stopping_criteria=[stopping_criteria],
|
| 359 |
)
|
|
@@ -385,7 +386,7 @@ def bot(history):
|
|
| 385 |
for upload_img in all_image_path:
|
| 386 |
api.upload_file(
|
| 387 |
path_or_fileobj=upload_img,
|
| 388 |
-
path_in_repo=upload_img.replace("./logs/"),
|
| 389 |
repo_id=repo_name,
|
| 390 |
repo_type="dataset",
|
| 391 |
# revision=revision,
|
|
@@ -394,7 +395,7 @@ def bot(history):
|
|
| 394 |
# upload json
|
| 395 |
api.upload_file(
|
| 396 |
path_or_fileobj=get_conv_log_filename(),
|
| 397 |
-
path_in_repo=get_conv_log_filename().replace("./logs/"),
|
| 398 |
repo_id=repo_name,
|
| 399 |
repo_type="dataset")
|
| 400 |
|
|
@@ -416,6 +417,31 @@ with gr.Blocks(
|
|
| 416 |
gr.HTML(html_header)
|
| 417 |
|
| 418 |
with gr.Column():
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 419 |
with gr.Row():
|
| 420 |
chatbot = gr.Chatbot([], elem_id="Pangea", bubble_full_width=False, height=750)
|
| 421 |
|
|
@@ -608,7 +634,7 @@ with gr.Blocks(
|
|
| 608 |
chat_msg = chat_input.submit(
|
| 609 |
add_message, [chatbot, chat_input], [chatbot, chat_input]
|
| 610 |
)
|
| 611 |
-
bot_msg = chat_msg.then(bot, chatbot, chatbot, api_name="bot_response")
|
| 612 |
bot_msg.then(lambda: gr.MultimodalTextbox(interactive=True), None, [chat_input])
|
| 613 |
|
| 614 |
# chatbot.like(print_like_dislike, None, None)
|
|
|
|
| 54 |
from huggingface_hub import login
|
| 55 |
from huggingface_hub import revision_exists
|
| 56 |
|
| 57 |
+
|
| 58 |
api = HfApi()
|
| 59 |
repo_name = "paralym/pangea_logs"
|
| 60 |
|
|
|
|
| 214 |
|
| 215 |
|
| 216 |
@spaces.GPU
|
| 217 |
+
def bot(history, temperature, top_p, max_output_tokens):
|
| 218 |
print("### turn start history",history)
|
| 219 |
print("### turn start conv",our_chatbot.conversation)
|
| 220 |
text = history[-1][0]
|
|
|
|
| 351 |
inputs=input_ids,
|
| 352 |
streamer=streamer,
|
| 353 |
images=image_tensor,
|
|
|
|
| 354 |
do_sample=True,
|
| 355 |
+
temperature=temperature,
|
| 356 |
+
top_p=top_p,
|
| 357 |
+
max_new_tokens=max_output_tokens,
|
| 358 |
use_cache=False,
|
| 359 |
stopping_criteria=[stopping_criteria],
|
| 360 |
)
|
|
|
|
| 386 |
for upload_img in all_image_path:
|
| 387 |
api.upload_file(
|
| 388 |
path_or_fileobj=upload_img,
|
| 389 |
+
path_in_repo=upload_img.replace("./logs/", ""),
|
| 390 |
repo_id=repo_name,
|
| 391 |
repo_type="dataset",
|
| 392 |
# revision=revision,
|
|
|
|
| 395 |
# upload json
|
| 396 |
api.upload_file(
|
| 397 |
path_or_fileobj=get_conv_log_filename(),
|
| 398 |
+
path_in_repo=get_conv_log_filename().replace("./logs/", ""),
|
| 399 |
repo_id=repo_name,
|
| 400 |
repo_type="dataset")
|
| 401 |
|
|
|
|
| 417 |
gr.HTML(html_header)
|
| 418 |
|
| 419 |
with gr.Column():
|
| 420 |
+
with gr.Accordion("Parameters", open=False) as parameter_row:
|
| 421 |
+
temperature = gr.Slider(
|
| 422 |
+
minimum=0.0,
|
| 423 |
+
maximum=1.0,
|
| 424 |
+
value=0.7,
|
| 425 |
+
step=0.1,
|
| 426 |
+
interactive=True,
|
| 427 |
+
label="Temperature",
|
| 428 |
+
)
|
| 429 |
+
top_p = gr.Slider(
|
| 430 |
+
minimum=0.0,
|
| 431 |
+
maximum=1.0,
|
| 432 |
+
value=1,
|
| 433 |
+
step=0.1,
|
| 434 |
+
interactive=True,
|
| 435 |
+
label="Top P",
|
| 436 |
+
)
|
| 437 |
+
max_output_tokens = gr.Slider(
|
| 438 |
+
minimum=0,
|
| 439 |
+
maximum=8192,
|
| 440 |
+
value=4096,
|
| 441 |
+
step=256,
|
| 442 |
+
interactive=True,
|
| 443 |
+
label="Max output tokens",
|
| 444 |
+
)
|
| 445 |
with gr.Row():
|
| 446 |
chatbot = gr.Chatbot([], elem_id="Pangea", bubble_full_width=False, height=750)
|
| 447 |
|
|
|
|
| 634 |
chat_msg = chat_input.submit(
|
| 635 |
add_message, [chatbot, chat_input], [chatbot, chat_input]
|
| 636 |
)
|
| 637 |
+
bot_msg = chat_msg.then(bot, [chatbot,temperature, top_p, max_output_tokens], chatbot, api_name="bot_response")
|
| 638 |
bot_msg.then(lambda: gr.MultimodalTextbox(interactive=True), None, [chat_input])
|
| 639 |
|
| 640 |
# chatbot.like(print_like_dislike, None, None)
|