prithivMLmods commited on
Commit
57d0261
·
verified ·
1 Parent(s): bcfb01b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -18,7 +18,7 @@ from qwen_vl_utils import process_vision_info
18
 
19
  # Constants for text generation
20
  MAX_MAX_NEW_TOKENS = 4096
21
- DEFAULT_MAX_NEW_TOKENS = 2048
22
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
23
 
24
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
@@ -312,4 +312,4 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
312
  outputs=[output, markdown_output])
313
 
314
  if __name__ == "__main__":
315
- demo.queue(max_size=30).launch(share=True, ssr_mode=False, show_error=True)
 
18
 
19
  # Constants for text generation
20
  MAX_MAX_NEW_TOKENS = 4096
21
+ DEFAULT_MAX_NEW_TOKENS = 3000
22
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
23
 
24
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
 
312
  outputs=[output, markdown_output])
313
 
314
  if __name__ == "__main__":
315
+ demo.queue(max_size=30).launch(share=True, mcp_server=True, ssr_mode=False, show_error=True)