Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
import gradio as gr
|
| 2 |
import torch
|
| 3 |
import time
|
| 4 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
|
@@ -79,7 +79,7 @@ def generate_response(message, temperature, max_new_tokens, top_p, task):
|
|
| 79 |
print(f"Prompt: {prompt}")
|
| 80 |
start_time = time.time()
|
| 81 |
inputs = tokenizer.encode(prompt, return_tensors="pt").to(model.device)
|
| 82 |
-
outputs = model.generate(input_ids=inputs,
|
| 83 |
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 84 |
processed_response = response.split("RESPONSE :")[-1].strip()
|
| 85 |
end_time = time.time()
|
|
@@ -111,4 +111,4 @@ with gr.Blocks(theme='1024m/1024m-1') as demo:
|
|
| 111 |
send_btn.click(fn=generate_response, inputs=[input_text, temperature, max_new_tokens, top_p, task_dropdown], outputs=output_text)
|
| 112 |
clear_btn.click(fn=lambda: ("", ""), inputs=None, outputs=[input_text, output_text])
|
| 113 |
if __name__ == "__main__":
|
| 114 |
-
demo.queue().launch()
|
|
|
|
| 1 |
+
"""import gradio as gr
|
| 2 |
import torch
|
| 3 |
import time
|
| 4 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
|
|
|
| 79 |
print(f"Prompt: {prompt}")
|
| 80 |
start_time = time.time()
|
| 81 |
inputs = tokenizer.encode(prompt, return_tensors="pt").to(model.device)
|
| 82 |
+
outputs = model.generate(input_ids=inputs, max_new_tokens=max_new_tokens, use_cache=True, temperature=temperature, min_p=top_p, pad_token_id=tokenizer.eos_token_id)
|
| 83 |
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 84 |
processed_response = response.split("RESPONSE :")[-1].strip()
|
| 85 |
end_time = time.time()
|
|
|
|
| 111 |
send_btn.click(fn=generate_response, inputs=[input_text, temperature, max_new_tokens, top_p, task_dropdown], outputs=output_text)
|
| 112 |
clear_btn.click(fn=lambda: ("", ""), inputs=None, outputs=[input_text, output_text])
|
| 113 |
if __name__ == "__main__":
|
| 114 |
+
demo.queue().launch()
|