Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -18,7 +18,7 @@ from qwen_vl_utils import process_vision_info
|
|
18 |
|
19 |
# Constants for text generation
|
20 |
MAX_MAX_NEW_TOKENS = 4096
|
21 |
-
DEFAULT_MAX_NEW_TOKENS =
|
22 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
23 |
|
24 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
@@ -312,4 +312,4 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
|
312 |
outputs=[output, markdown_output])
|
313 |
|
314 |
if __name__ == "__main__":
|
315 |
-
demo.queue(max_size=30).launch(share=True, ssr_mode=False, show_error=True)
|
|
|
18 |
|
19 |
# Constants for text generation
|
20 |
MAX_MAX_NEW_TOKENS = 4096
|
21 |
+
DEFAULT_MAX_NEW_TOKENS = 3000
|
22 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
23 |
|
24 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
|
|
312 |
outputs=[output, markdown_output])
|
313 |
|
314 |
if __name__ == "__main__":
|
315 |
+
demo.queue(max_size=30).launch(share=True, mcp_server=True, ssr_mode=False, show_error=True)
|