Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -38,9 +38,9 @@ def create_chat_template_messages(history, prompt):
|
|
| 38 |
@spaces.GPU
|
| 39 |
def generate_responses(prompt, history):
|
| 40 |
messages = create_chat_template_messages(history, prompt)
|
| 41 |
-
wrapped_prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_special_tokens=
|
| 42 |
|
| 43 |
-
inputs = tokenizer.encode(wrapped_prompt, add_special_tokens=
|
| 44 |
|
| 45 |
async def custom_sampler_task():
|
| 46 |
generated_list = []
|
|
|
|
| 38 |
@spaces.GPU
|
| 39 |
def generate_responses(prompt, history):
|
| 40 |
messages = create_chat_template_messages(history, prompt)
|
| 41 |
+
wrapped_prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_special_tokens=False, add_generation_prompt=True)
|
| 42 |
|
| 43 |
+
inputs = tokenizer.encode(wrapped_prompt, add_special_tokens=True, return_tensors="pt").to("cuda")
|
| 44 |
|
| 45 |
async def custom_sampler_task():
|
| 46 |
generated_list = []
|