prithivMLmods commited on
Commit
e75c6cf
·
verified ·
1 Parent(s): ff5cc66

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -4
app.py CHANGED
@@ -21,13 +21,12 @@ from transformers import (
21
  from transformers.image_utils import load_image
22
 
23
  # Constants for text generation
24
- MAX_MAX_NEW_TOKENS = 2048
25
- DEFAULT_MAX_NEW_TOKENS = 1024
26
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
27
 
28
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
29
 
30
- # Load multimodal processor and model (Callisto OCR3)
31
  MODEL_ID = "nvidia/Cosmos-Reason1-7B"
32
  processor = AutoProcessor.from_pretrained(MODEL_ID, trust_remote_code=True)
33
  model_m = Qwen2_5_VLForConditionalGeneration.from_pretrained(
@@ -207,4 +206,4 @@ with gr.Blocks() as demo:
207
  )
208
 
209
  if __name__ == "__main__":
210
- demo.queue(max_size=20).launch(share=True)
 
21
  from transformers.image_utils import load_image
22
 
23
  # Constants for text generation
24
+ MAX_MAX_NEW_TOKENS = 4096
25
+ DEFAULT_MAX_NEW_TOKENS = 1536
26
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
27
 
28
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
29
 
 
30
  MODEL_ID = "nvidia/Cosmos-Reason1-7B"
31
  processor = AutoProcessor.from_pretrained(MODEL_ID, trust_remote_code=True)
32
  model_m = Qwen2_5_VLForConditionalGeneration.from_pretrained(
 
206
  )
207
 
208
  if __name__ == "__main__":
209
+ demo.queue(max_size=30).launch(share=True, ssr_mode=False, show_error=True)