Csplk commited on
Commit
733bfde
ยท
verified ยท
1 Parent(s): 2754d68

Update app.py

Browse files

Testing new model and gpu hold on !

Files changed (1) hide show
  1. app.py +11 -0
app.py CHANGED
@@ -10,6 +10,15 @@ from torchvision.transforms.v2 import Resize
10
  import subprocess
11
  subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
12
 
 
 
 
 
 
 
 
 
 
13
  model_id = "vikhyatk/moondream2"
14
  revision = "2024-08-26"
15
  tokenizer = AutoTokenizer.from_pretrained(model_id, revision=revision)
@@ -18,6 +27,8 @@ moondream = AutoModelForCausalLM.from_pretrained(
18
  torch_dtype=torch.bfloat16, device_map={"": "cuda"},
19
  attn_implementation="flash_attention_2"
20
  )
 
 
21
  moondream.eval()
22
 
23
  @spaces.GPU
 
10
  import subprocess
11
  subprocess.run('pip install flash-attn --no-build-isolation', env={'FLASH_ATTENTION_SKIP_CUDA_BUILD': "TRUE"}, shell=True)
12
 
13
+ mooondream = AutoModelForCausalLM.from_pretrained(
14
+ "vikhyatk/moondream2",
15
+ revision="2025-01-09",
16
+ trust_remote_code=True,
17
+ device_map={"": "cuda"},
18
+ #attn_implementation="flash_attention_2"
19
+ )
20
+
21
+ '''
22
  model_id = "vikhyatk/moondream2"
23
  revision = "2024-08-26"
24
  tokenizer = AutoTokenizer.from_pretrained(model_id, revision=revision)
 
27
  torch_dtype=torch.bfloat16, device_map={"": "cuda"},
28
  attn_implementation="flash_attention_2"
29
  )
30
+ '''
31
+
32
  moondream.eval()
33
 
34
  @spaces.GPU