Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -170,36 +170,43 @@ def send_to_model(*args, **kwargs):
|
|
| 170 |
|
| 171 |
def send_to_model_impl(prompt, model_selection, hf_model_choice, hf_custom_model, hf_api_key,
|
| 172 |
groq_model_choice, groq_api_key, openai_api_key):
|
| 173 |
-
|
| 174 |
-
|
| 175 |
-
|
| 176 |
-
|
| 177 |
-
|
| 178 |
-
|
| 179 |
-
|
| 180 |
-
|
| 181 |
-
|
| 182 |
-
|
| 183 |
-
|
| 184 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 185 |
|
| 186 |
-
|
| 187 |
-
|
| 188 |
-
|
| 189 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 190 |
|
| 191 |
-
else:
|
| 192 |
-
return "Invalid model selection.", []
|
| 193 |
-
|
| 194 |
-
if summary.startswith("Error"):
|
| 195 |
-
return summary, []
|
| 196 |
-
|
| 197 |
-
# Save summary for download
|
| 198 |
-
with tempfile.NamedTemporaryFile(delete=False, mode='w', suffix='.txt') as f:
|
| 199 |
-
f.write(summary)
|
| 200 |
-
|
| 201 |
-
return summary, [f.name]
|
| 202 |
-
|
| 203 |
def send_to_hf_inference(prompt: str, model_name: str, api_key: str) -> str:
|
| 204 |
"""Send prompt to HuggingFace using Inference API"""
|
| 205 |
try:
|
|
@@ -266,21 +273,17 @@ def send_to_openai(prompt: str, api_key: str) -> str:
|
|
| 266 |
return f"Error with OpenAI API: {e}"
|
| 267 |
|
| 268 |
def copy_text_js(element_id: str) -> str:
|
| 269 |
-
return f"""
|
| 270 |
-
|
| 271 |
-
|
| 272 |
-
|
| 273 |
-
|
| 274 |
-
|
| 275 |
-
|
| 276 |
-
|
| 277 |
-
|
| 278 |
-
}} catch (e) {{
|
| 279 |
-
console.error(e);
|
| 280 |
-
return "Failed to copy: " + e.message;
|
| 281 |
-
}}
|
| 282 |
}}
|
| 283 |
-
"""
|
| 284 |
|
| 285 |
def open_chatgpt() -> str:
|
| 286 |
"""Open ChatGPT in new browser tab"""
|
|
@@ -354,44 +357,16 @@ def generate_prompt(text, template, snippet_idx=None):
|
|
| 354 |
logging.error(f"Error generating prompt: {e}")
|
| 355 |
return f"Error generating prompt: {str(e)}", "", None
|
| 356 |
|
| 357 |
-
def download_file(content: str, prefix: str = "file") -> List[str]:
|
| 358 |
-
"""Create a downloadable file with content and better error handling"""
|
| 359 |
-
if not content:
|
| 360 |
-
return []
|
| 361 |
-
try:
|
| 362 |
-
with tempfile.NamedTemporaryFile(delete=False, mode='w', suffix='.txt', prefix=prefix) as f:
|
| 363 |
-
f.write(content)
|
| 364 |
-
return [f.name]
|
| 365 |
-
except Exception as e:
|
| 366 |
-
logging.error(f"Error creating download file: {e}")
|
| 367 |
-
return []
|
| 368 |
-
|
| 369 |
# Main Interface
|
| 370 |
-
with gr.Blocks(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 371 |
# State variables
|
| 372 |
pdf_content = gr.State("")
|
| 373 |
snippets = gr.State([])
|
| 374 |
-
|
| 375 |
-
# CSS for responsiveness
|
| 376 |
-
demo.css = """
|
| 377 |
-
.gradio-container {
|
| 378 |
-
max-width: 90%; /* Adjust as needed */
|
| 379 |
-
margin: 0 auto; /* Center the container */
|
| 380 |
-
}
|
| 381 |
-
@media (max-width: 768px) { /* Example breakpoint for smaller screens */
|
| 382 |
-
.gradio-container {
|
| 383 |
-
max-width: 98%;
|
| 384 |
-
padding: 10px; /* Add padding for better mobile experience */
|
| 385 |
-
}
|
| 386 |
-
.gr-row {
|
| 387 |
-
flex-direction: column; /* Stack elements vertically */
|
| 388 |
-
}
|
| 389 |
-
.gr-col {
|
| 390 |
-
width: 100%; /* Make columns full width */
|
| 391 |
-
margin-bottom: 10px; /* Add spacing between columns */
|
| 392 |
-
}
|
| 393 |
-
}
|
| 394 |
-
"""
|
| 395 |
|
| 396 |
# Header
|
| 397 |
gr.Markdown("# π Smart PDF Summarizer")
|
|
@@ -421,14 +396,15 @@ with gr.Blocks(theme=gr.themes.Default()) as demo:
|
|
| 421 |
label="Context Size"
|
| 422 |
)
|
| 423 |
|
|
|
|
| 424 |
with gr.Row():
|
| 425 |
for size_name, size_value in CONTEXT_SIZES.items():
|
| 426 |
gr.Button(
|
| 427 |
-
size_name,
|
| 428 |
size="sm",
|
| 429 |
scale=1
|
| 430 |
).click(
|
| 431 |
-
lambda v=size_value: v,
|
| 432 |
None,
|
| 433 |
context_size
|
| 434 |
)
|
|
@@ -487,10 +463,10 @@ with gr.Blocks(theme=gr.themes.Default()) as demo:
|
|
| 487 |
with gr.Row():
|
| 488 |
with gr.Column(scale=1):
|
| 489 |
model_choice = gr.Radio(
|
| 490 |
-
|
| 491 |
-
|
| 492 |
-
|
| 493 |
-
|
| 494 |
|
| 495 |
with gr.Column(visible=False) as openai_options:
|
| 496 |
openai_api_key = gr.Textbox(
|
|
@@ -581,8 +557,15 @@ with gr.Blocks(theme=gr.themes.Default()) as demo:
|
|
| 581 |
)
|
| 582 |
|
| 583 |
def refresh_groq_models_list():
|
| 584 |
-
|
| 585 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 586 |
|
| 587 |
def toggle_custom_model(model_name):
|
| 588 |
return gr.update(visible=model_name == "Custom Model")
|
|
@@ -601,14 +584,12 @@ with gr.Blocks(theme=gr.themes.Default()) as demo:
|
|
| 601 |
return update_context_size("Groq API", model_name)
|
| 602 |
|
| 603 |
def handle_model_selection(choice):
|
| 604 |
-
|
| 605 |
-
|
| 606 |
-
|
| 607 |
-
gr.update(visible=choice == "
|
| 608 |
-
gr.update(
|
| 609 |
-
|
| 610 |
-
gr.update(value=ctx_size) # context_size
|
| 611 |
-
)
|
| 612 |
|
| 613 |
# PDF Processing Handlers
|
| 614 |
def handle_pdf_process(pdf, fmt, ctx_size):
|
|
@@ -699,24 +680,24 @@ with gr.Blocks(theme=gr.themes.Default()) as demo:
|
|
| 699 |
|
| 700 |
# Copy button handlers
|
| 701 |
def handle_prompt_generation(snippet_text, template, snippet_choice, snippets_list):
|
| 702 |
-
"""Generate prompt from selected snippet"""
|
| 703 |
-
if not snippet_text or not snippets_list:
|
| 704 |
-
return "No text available for prompt generation.", "", None
|
| 705 |
-
|
| 706 |
try:
|
|
|
|
|
|
|
|
|
|
| 707 |
idx = get_snippet_index(snippet_choice)
|
| 708 |
-
|
|
|
|
| 709 |
|
| 710 |
-
|
|
|
|
|
|
|
| 711 |
with tempfile.NamedTemporaryFile(delete=False, mode='w', suffix='.txt') as f:
|
| 712 |
f.write(prompt)
|
| 713 |
|
| 714 |
-
return "Prompt generated
|
| 715 |
-
|
| 716 |
except Exception as e:
|
| 717 |
-
|
| 718 |
-
|
| 719 |
-
return error_msg, "", None
|
| 720 |
|
| 721 |
def handle_copy_action(text):
|
| 722 |
"""Handle copy to clipboard action"""
|
|
@@ -784,42 +765,29 @@ with gr.Blocks(theme=gr.themes.Default()) as demo:
|
|
| 784 |
outputs=[context_size]
|
| 785 |
)
|
| 786 |
|
| 787 |
-
#
|
| 788 |
-
for
|
| 789 |
-
|
| 790 |
-
|
| 791 |
-
|
| 792 |
-
|
| 793 |
-
).click(
|
| 794 |
-
lambda v=size_value: gr.update(value=v),
|
| 795 |
-
None,
|
| 796 |
-
context_size
|
| 797 |
-
)
|
| 798 |
-
|
| 799 |
-
# Download handlers (simplified)
|
| 800 |
-
for btn, content in [
|
| 801 |
-
(download_full_text, pdf_content),
|
| 802 |
-
(download_snippet, generated_prompt),
|
| 803 |
-
(download_prompt, generated_prompt),
|
| 804 |
-
(download_summary, summary_output)
|
| 805 |
]:
|
| 806 |
btn.click(
|
| 807 |
-
lambda x:
|
| 808 |
inputs=[content],
|
| 809 |
outputs=[download_files]
|
| 810 |
)
|
| 811 |
|
| 812 |
-
|
| 813 |
-
|
| 814 |
-
|
| 815 |
-
|
| 816 |
-
|
| 817 |
-
|
| 818 |
-
|
| 819 |
-
|
| 820 |
-
|
| 821 |
-
|
| 822 |
-
)
|
| 823 |
|
| 824 |
# ChatGPT handler
|
| 825 |
open_chatgpt_button.click(
|
|
|
|
| 170 |
|
| 171 |
def send_to_model_impl(prompt, model_selection, hf_model_choice, hf_custom_model, hf_api_key,
|
| 172 |
groq_model_choice, groq_api_key, openai_api_key):
|
| 173 |
+
try:
|
| 174 |
+
if model_selection == "Clipboard only":
|
| 175 |
+
return "Use copy/paste for processing", []
|
| 176 |
+
|
| 177 |
+
if model_selection == "HuggingFace Inference":
|
| 178 |
+
if not hf_api_key:
|
| 179 |
+
return "Error: HuggingFace API key required", []
|
| 180 |
+
if not hf_model_choice:
|
| 181 |
+
return "Error: Select a HuggingFace model", []
|
| 182 |
+
model_id = hf_custom_model if hf_model_choice == "Custom Model" else model_registry.hf_models[hf_model_choice]
|
| 183 |
+
summary = send_to_hf_inference(prompt, model_id, hf_api_key)
|
| 184 |
+
|
| 185 |
+
elif model_selection == "Groq API":
|
| 186 |
+
if not groq_api_key:
|
| 187 |
+
return "Error: Groq API key required", []
|
| 188 |
+
if not groq_model_choice:
|
| 189 |
+
return "Error: Select a Groq model", []
|
| 190 |
+
summary = send_to_groq(prompt, groq_model_choice, groq_api_key)
|
| 191 |
+
|
| 192 |
+
elif model_selection == "OpenAI ChatGPT":
|
| 193 |
+
if not openai_api_key:
|
| 194 |
+
return "Error: OpenAI API key required", []
|
| 195 |
+
summary = send_to_openai(prompt, openai_api_key)
|
| 196 |
+
|
| 197 |
+
else:
|
| 198 |
+
return "Error: Invalid model selection", []
|
| 199 |
|
| 200 |
+
# Save summary for download
|
| 201 |
+
with tempfile.NamedTemporaryFile(delete=False, mode='w', suffix='.txt') as f:
|
| 202 |
+
f.write(summary)
|
| 203 |
+
|
| 204 |
+
return summary, [f.name]
|
| 205 |
+
except Exception as e:
|
| 206 |
+
error_msg = f"Error processing request: {str(e)}"
|
| 207 |
+
logging.error(error_msg)
|
| 208 |
+
return error_msg, []
|
| 209 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 210 |
def send_to_hf_inference(prompt: str, model_name: str, api_key: str) -> str:
|
| 211 |
"""Send prompt to HuggingFace using Inference API"""
|
| 212 |
try:
|
|
|
|
| 273 |
return f"Error with OpenAI API: {e}"
|
| 274 |
|
| 275 |
def copy_text_js(element_id: str) -> str:
|
| 276 |
+
return f"""function() {{
|
| 277 |
+
let textarea = document.getElementById('{element_id}');
|
| 278 |
+
if (!textarea) return 'Element not found';
|
| 279 |
+
textarea.select();
|
| 280 |
+
try {{
|
| 281 |
+
document.execCommand('copy');
|
| 282 |
+
return 'Copied to clipboard!';
|
| 283 |
+
}} catch(err) {{
|
| 284 |
+
return 'Failed to copy: ' + err;
|
|
|
|
|
|
|
|
|
|
|
|
|
| 285 |
}}
|
| 286 |
+
}}"""
|
| 287 |
|
| 288 |
def open_chatgpt() -> str:
|
| 289 |
"""Open ChatGPT in new browser tab"""
|
|
|
|
| 357 |
logging.error(f"Error generating prompt: {e}")
|
| 358 |
return f"Error generating prompt: {str(e)}", "", None
|
| 359 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 360 |
# Main Interface
|
| 361 |
+
with gr.Blocks(css="""
|
| 362 |
+
.gradio-container {max-width: 90%; margin: 0 auto;}
|
| 363 |
+
@media (max-width: 768px) {.gradio-container {max-width: 98%; padding: 10px;} .gr-row {flex-direction: column;} .gr-col {width: 100%; margin-bottom: 10px;}}
|
| 364 |
+
""") as demo:
|
| 365 |
+
gr.Markdown("# π Smart PDF Summarizer")
|
| 366 |
+
|
| 367 |
# State variables
|
| 368 |
pdf_content = gr.State("")
|
| 369 |
snippets = gr.State([])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 370 |
|
| 371 |
# Header
|
| 372 |
gr.Markdown("# π Smart PDF Summarizer")
|
|
|
|
| 396 |
label="Context Size"
|
| 397 |
)
|
| 398 |
|
| 399 |
+
gr.Markdown("### Context Size")
|
| 400 |
with gr.Row():
|
| 401 |
for size_name, size_value in CONTEXT_SIZES.items():
|
| 402 |
gr.Button(
|
| 403 |
+
size_name,
|
| 404 |
size="sm",
|
| 405 |
scale=1
|
| 406 |
).click(
|
| 407 |
+
lambda v=size_value: gr.update(value=v),
|
| 408 |
None,
|
| 409 |
context_size
|
| 410 |
)
|
|
|
|
| 463 |
with gr.Row():
|
| 464 |
with gr.Column(scale=1):
|
| 465 |
model_choice = gr.Radio(
|
| 466 |
+
choices=["Clipboard only", "OpenAI ChatGPT", "HuggingFace Inference", "Groq API"],
|
| 467 |
+
value="Clipboard only",
|
| 468 |
+
label="π€ Model Selection"
|
| 469 |
+
)
|
| 470 |
|
| 471 |
with gr.Column(visible=False) as openai_options:
|
| 472 |
openai_api_key = gr.Textbox(
|
|
|
|
| 557 |
)
|
| 558 |
|
| 559 |
def refresh_groq_models_list():
|
| 560 |
+
try:
|
| 561 |
+
with gr.Progress() as progress:
|
| 562 |
+
progress(0, "Refreshing Groq models...")
|
| 563 |
+
updated_models = model_registry.refresh_groq_models()
|
| 564 |
+
progress(1, "Complete!")
|
| 565 |
+
return gr.update(choices=list(updated_models.keys()))
|
| 566 |
+
except Exception as e:
|
| 567 |
+
logging.error(f"Error refreshing models: {e}")
|
| 568 |
+
return gr.update()
|
| 569 |
|
| 570 |
def toggle_custom_model(model_name):
|
| 571 |
return gr.update(visible=model_name == "Custom Model")
|
|
|
|
| 584 |
return update_context_size("Groq API", model_name)
|
| 585 |
|
| 586 |
def handle_model_selection(choice):
|
| 587 |
+
return [
|
| 588 |
+
gr.update(visible=choice == "HuggingFace Inference"),
|
| 589 |
+
gr.update(visible=choice == "Groq API"),
|
| 590 |
+
gr.update(visible=choice == "OpenAI ChatGPT"),
|
| 591 |
+
gr.update(value=get_model_context_size(choice))
|
| 592 |
+
]
|
|
|
|
|
|
|
| 593 |
|
| 594 |
# PDF Processing Handlers
|
| 595 |
def handle_pdf_process(pdf, fmt, ctx_size):
|
|
|
|
| 680 |
|
| 681 |
# Copy button handlers
|
| 682 |
def handle_prompt_generation(snippet_text, template, snippet_choice, snippets_list):
|
|
|
|
|
|
|
|
|
|
|
|
|
| 683 |
try:
|
| 684 |
+
if not snippets_list:
|
| 685 |
+
return "No text available.", "", None
|
| 686 |
+
|
| 687 |
idx = get_snippet_index(snippet_choice)
|
| 688 |
+
base_prompt = template if template else "Summarize the following text:"
|
| 689 |
+
content = snippets_list[idx]
|
| 690 |
|
| 691 |
+
prompt = f"{base_prompt}\n---\n{content}\n---"
|
| 692 |
+
|
| 693 |
+
# Save prompt for download
|
| 694 |
with tempfile.NamedTemporaryFile(delete=False, mode='w', suffix='.txt') as f:
|
| 695 |
f.write(prompt)
|
| 696 |
|
| 697 |
+
return "Prompt generated!", prompt, [f.name]
|
|
|
|
| 698 |
except Exception as e:
|
| 699 |
+
logging.error(f"Error generating prompt: {e}")
|
| 700 |
+
return f"Error: {str(e)}", "", None
|
|
|
|
| 701 |
|
| 702 |
def handle_copy_action(text):
|
| 703 |
"""Handle copy to clipboard action"""
|
|
|
|
| 765 |
outputs=[context_size]
|
| 766 |
)
|
| 767 |
|
| 768 |
+
# Download handlers
|
| 769 |
+
for btn, content, prefix in [
|
| 770 |
+
(download_full_text, pdf_content, "full_text"),
|
| 771 |
+
(download_snippet, generated_prompt, "snippet"),
|
| 772 |
+
(download_prompt, generated_prompt, "prompt"),
|
| 773 |
+
(download_summary, summary_output, "summary")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 774 |
]:
|
| 775 |
btn.click(
|
| 776 |
+
lambda x, p=prefix: download_file(x, p) if x else [],
|
| 777 |
inputs=[content],
|
| 778 |
outputs=[download_files]
|
| 779 |
)
|
| 780 |
|
| 781 |
+
def download_file(content: str, prefix: str) -> List[str]:
|
| 782 |
+
if not content:
|
| 783 |
+
return []
|
| 784 |
+
try:
|
| 785 |
+
with tempfile.NamedTemporaryFile(delete=False, mode='w', suffix='.txt', prefix=prefix) as f:
|
| 786 |
+
f.write(content)
|
| 787 |
+
return [f.name]
|
| 788 |
+
except Exception as e:
|
| 789 |
+
logging.error(f"Error creating download file: {e}")
|
| 790 |
+
return []
|
|
|
|
| 791 |
|
| 792 |
# ChatGPT handler
|
| 793 |
open_chatgpt_button.click(
|