Update modeling_gpt2vision.py
Browse files- modeling_gpt2vision.py +2 -12
modeling_gpt2vision.py
CHANGED
@@ -49,7 +49,7 @@ class GPT2Vision(PreTrainedModel):
|
|
49 |
self.tokenizer = tokenizer
|
50 |
tokenizer.pad_token = tokenizer.eos_token
|
51 |
self.image_token_id = self.tokenizer.convert_tokens_to_ids(IMAGE_TOKEN)
|
52 |
-
self.img_tokens = 197
|
53 |
|
54 |
@property
|
55 |
def device(self):
|
@@ -61,12 +61,11 @@ class GPT2Vision(PreTrainedModel):
|
|
61 |
if isinstance(text, str):
|
62 |
text = [text]
|
63 |
|
64 |
-
print("text",text)
|
65 |
text_inputs = self.tokenizer(
|
66 |
text,
|
67 |
padding='max_length',
|
68 |
truncation=True,
|
69 |
-
max_length=
|
70 |
return_tensors="pt",
|
71 |
).to(device)
|
72 |
|
@@ -97,15 +96,6 @@ class GPT2Vision(PreTrainedModel):
|
|
97 |
|
98 |
return inputs_embeds, attention_mask
|
99 |
|
100 |
-
def forward(self, batch, **kwargs):
|
101 |
-
inputs_embeds, attention_mask = self.preprocess_inputs(batch)
|
102 |
-
|
103 |
-
outputs = self.language_model(
|
104 |
-
inputs_embeds=inputs_embeds,
|
105 |
-
attention_mask=attention_mask,
|
106 |
-
**kwargs
|
107 |
-
)
|
108 |
-
return outputs
|
109 |
|
110 |
def generate(self, question, image, max_new_tokens=30, **kwargs):
|
111 |
prompt = f"\n\nQuestion: {question}\n\nAnswer:"
|
|
|
49 |
self.tokenizer = tokenizer
|
50 |
tokenizer.pad_token = tokenizer.eos_token
|
51 |
self.image_token_id = self.tokenizer.convert_tokens_to_ids(IMAGE_TOKEN)
|
52 |
+
self.img_tokens = 197
|
53 |
|
54 |
@property
|
55 |
def device(self):
|
|
|
61 |
if isinstance(text, str):
|
62 |
text = [text]
|
63 |
|
|
|
64 |
text_inputs = self.tokenizer(
|
65 |
text,
|
66 |
padding='max_length',
|
67 |
truncation=True,
|
68 |
+
max_length=384,
|
69 |
return_tensors="pt",
|
70 |
).to(device)
|
71 |
|
|
|
96 |
|
97 |
return inputs_embeds, attention_mask
|
98 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
99 |
|
100 |
def generate(self, question, image, max_new_tokens=30, **kwargs):
|
101 |
prompt = f"\n\nQuestion: {question}\n\nAnswer:"
|