Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -101,8 +101,8 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
|
|
| 101 |
"""
|
| 102 |
|
| 103 |
#Disable Qwen3 thinking
|
| 104 |
-
|
| 105 |
-
|
| 106 |
|
| 107 |
# The 'history' variable from Gradio contains the entire conversation.
|
| 108 |
# We prepend the system prompt to this history to form the final payload.
|
|
@@ -116,8 +116,7 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
|
|
| 116 |
response_stream = ollama.chat(
|
| 117 |
model=current_selected_model, # Use the dynamically selected model
|
| 118 |
messages=messages,
|
| 119 |
-
stream=True
|
| 120 |
-
think=False
|
| 121 |
)
|
| 122 |
|
| 123 |
# Iterate through the stream, updating the placeholder with each new chunk.
|
|
|
|
| 101 |
"""
|
| 102 |
|
| 103 |
#Disable Qwen3 thinking
|
| 104 |
+
if "Qwen3".lower() in current_selected_model:
|
| 105 |
+
system_prompt = system_prompt+" /no_think"
|
| 106 |
|
| 107 |
# The 'history' variable from Gradio contains the entire conversation.
|
| 108 |
# We prepend the system prompt to this history to form the final payload.
|
|
|
|
| 116 |
response_stream = ollama.chat(
|
| 117 |
model=current_selected_model, # Use the dynamically selected model
|
| 118 |
messages=messages,
|
| 119 |
+
stream=True
|
|
|
|
| 120 |
)
|
| 121 |
|
| 122 |
# Iterate through the stream, updating the placeholder with each new chunk.
|