Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -5,15 +5,17 @@ import open_clip
|
|
5 |
|
6 |
from huggingface_hub import hf_hub_download
|
7 |
|
|
|
|
|
8 |
torch.hub.download_url_to_file('http://images.cocodataset.org/val2017/000000039769.jpg', 'cats.jpg')
|
9 |
torch.hub.download_url_to_file('https://huggingface.co/datasets/nielsr/textcaps-sample/resolve/main/stop_sign.png', 'stop_sign.png')
|
10 |
torch.hub.download_url_to_file('https://cdn.openai.com/dall-e-2/demos/text2im/astronaut/horse/photo/0.jpg', 'astronaut.jpg')
|
11 |
|
12 |
git_processor_large_coco = AutoProcessor.from_pretrained("microsoft/git-large-coco")
|
13 |
-
git_model_large_coco = AutoModelForCausalLM.from_pretrained("microsoft/git-large-coco"
|
14 |
|
15 |
blip_processor_large = AutoProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
|
16 |
-
blip_model_large = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large"
|
17 |
|
18 |
blip2_processor = AutoProcessor.from_pretrained("Salesforce/blip2-opt-6.7b")
|
19 |
blip2_model_4_bit = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-opt-6.7b", device_map="auto", load_in_4bit=True, torch_dtype=torch.float16)
|
@@ -21,8 +23,6 @@ blip2_model_4_bit = Blip2ForConditionalGeneration.from_pretrained("Salesforce/bl
|
|
21 |
instructblip_processor = AutoProcessor.from_pretrained("Salesforce/instructblip-vicuna-7b")
|
22 |
instructblip_model_4_bit = InstructBlipForConditionalGeneration.from_pretrained("Salesforce/instructblip-vicuna-7b", device_map="auto", load_in_4bit=True, torch_dtype=torch.float16)
|
23 |
|
24 |
-
device = "cuda" if torch.cuda.is_available() else "cpu"
|
25 |
-
|
26 |
def generate_caption(processor, model, image, tokenizer=None, use_float_16=False):
|
27 |
inputs = processor(images=image, return_tensors="pt").to(device)
|
28 |
|
|
|
5 |
|
6 |
from huggingface_hub import hf_hub_download
|
7 |
|
8 |
+
device = "cuda" if torch.cuda.is_available() else "cpu"
|
9 |
+
|
10 |
torch.hub.download_url_to_file('http://images.cocodataset.org/val2017/000000039769.jpg', 'cats.jpg')
|
11 |
torch.hub.download_url_to_file('https://huggingface.co/datasets/nielsr/textcaps-sample/resolve/main/stop_sign.png', 'stop_sign.png')
|
12 |
torch.hub.download_url_to_file('https://cdn.openai.com/dall-e-2/demos/text2im/astronaut/horse/photo/0.jpg', 'astronaut.jpg')
|
13 |
|
14 |
git_processor_large_coco = AutoProcessor.from_pretrained("microsoft/git-large-coco")
|
15 |
+
git_model_large_coco = AutoModelForCausalLM.from_pretrained("microsoft/git-large-coco").to(device)
|
16 |
|
17 |
blip_processor_large = AutoProcessor.from_pretrained("Salesforce/blip-image-captioning-large")
|
18 |
+
blip_model_large = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-large").to(device)
|
19 |
|
20 |
blip2_processor = AutoProcessor.from_pretrained("Salesforce/blip2-opt-6.7b")
|
21 |
blip2_model_4_bit = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-opt-6.7b", device_map="auto", load_in_4bit=True, torch_dtype=torch.float16)
|
|
|
23 |
instructblip_processor = AutoProcessor.from_pretrained("Salesforce/instructblip-vicuna-7b")
|
24 |
instructblip_model_4_bit = InstructBlipForConditionalGeneration.from_pretrained("Salesforce/instructblip-vicuna-7b", device_map="auto", load_in_4bit=True, torch_dtype=torch.float16)
|
25 |
|
|
|
|
|
26 |
def generate_caption(processor, model, image, tokenizer=None, use_float_16=False):
|
27 |
inputs = processor(images=image, return_tensors="pt").to(device)
|
28 |
|