Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -45,7 +45,6 @@ Few sentences of reasoning
|
|
| 45 |
"""
|
| 46 |
|
| 47 |
# --- Helper Functions ---
|
| 48 |
-
@spaces.GPU
|
| 49 |
def format_rules(rules):
|
| 50 |
formatted_rules = "<rules>\n"
|
| 51 |
for i, rule in enumerate(rules):
|
|
@@ -142,7 +141,8 @@ class ModelWrapper:
|
|
| 142 |
message, tokenize=False, continue_final_message=True, enable_thinking=False
|
| 143 |
)
|
| 144 |
return prompt
|
| 145 |
-
|
|
|
|
| 146 |
def get_response(self, input, temperature=0.7, top_k=20, top_p=0.8, max_new_tokens=256,
|
| 147 |
enable_thinking=True, system_prompt=SYSTEM_PROMPT):
|
| 148 |
print("Generating response...")
|
|
|
|
| 45 |
"""
|
| 46 |
|
| 47 |
# --- Helper Functions ---
|
|
|
|
| 48 |
def format_rules(rules):
|
| 49 |
formatted_rules = "<rules>\n"
|
| 50 |
for i, rule in enumerate(rules):
|
|
|
|
| 141 |
message, tokenize=False, continue_final_message=True, enable_thinking=False
|
| 142 |
)
|
| 143 |
return prompt
|
| 144 |
+
|
| 145 |
+
@spaces.GPU
|
| 146 |
def get_response(self, input, temperature=0.7, top_k=20, top_p=0.8, max_new_tokens=256,
|
| 147 |
enable_thinking=True, system_prompt=SYSTEM_PROMPT):
|
| 148 |
print("Generating response...")
|