Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -18,7 +18,7 @@ from qwen_vl_utils import process_vision_info
|
|
| 18 |
|
| 19 |
# Constants for text generation
|
| 20 |
MAX_MAX_NEW_TOKENS = 4096
|
| 21 |
-
DEFAULT_MAX_NEW_TOKENS =
|
| 22 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
| 23 |
|
| 24 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
|
@@ -312,4 +312,4 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
|
| 312 |
outputs=[output, markdown_output])
|
| 313 |
|
| 314 |
if __name__ == "__main__":
|
| 315 |
-
demo.queue(max_size=30).launch(share=True, ssr_mode=False, show_error=True)
|
|
|
|
| 18 |
|
| 19 |
# Constants for text generation
|
| 20 |
MAX_MAX_NEW_TOKENS = 4096
|
| 21 |
+
DEFAULT_MAX_NEW_TOKENS = 3000
|
| 22 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
| 23 |
|
| 24 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
|
|
|
| 312 |
outputs=[output, markdown_output])
|
| 313 |
|
| 314 |
if __name__ == "__main__":
|
| 315 |
+
demo.queue(max_size=30).launch(share=True, mcp_server=True, ssr_mode=False, show_error=True)
|