Spaces:
Running
Running
laylm added
Browse files
app.py
CHANGED
@@ -18,7 +18,7 @@ import librosa
|
|
18 |
from PIL import Image
|
19 |
from torch.nn.utils import rnn
|
20 |
from gtts import gTTS
|
21 |
-
|
22 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
23 |
|
24 |
class LabelClassifier(nn.Module):
|
@@ -125,15 +125,22 @@ def m3(que, image):
|
|
125 |
# sequence = re.sub(r"<.*?>", "", sequence, count=1).strip() # remove first task start token
|
126 |
# return processor3.token2json(sequence)['answer']
|
127 |
|
|
|
|
|
|
|
|
|
128 |
|
129 |
-
|
130 |
-
|
|
|
|
|
|
|
131 |
|
132 |
|
133 |
-
inputs = processor3(images=image, text=que, return_tensors="pt")
|
134 |
|
135 |
-
predictions = model3.generate(**inputs)
|
136 |
-
return processor3.decode(predictions[0], skip_special_tokens=True)
|
137 |
|
138 |
|
139 |
|
|
|
18 |
from PIL import Image
|
19 |
from torch.nn.utils import rnn
|
20 |
from gtts import gTTS
|
21 |
+
import layoutlm
|
22 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
23 |
|
24 |
class LabelClassifier(nn.Module):
|
|
|
125 |
# sequence = re.sub(r"<.*?>", "", sequence, count=1).strip() # remove first task start token
|
126 |
# return processor3.token2json(sequence)['answer']
|
127 |
|
128 |
+
result = ''
|
129 |
+
if image is None:
|
130 |
+
return "Please select an image", None
|
131 |
+
|
132 |
|
133 |
+
result = layoutlm.get_result(image, que)
|
134 |
+
return result
|
135 |
+
|
136 |
+
# model3 = Pix2StructForConditionalGeneration.from_pretrained("google/pix2struct-docvqa-large")
|
137 |
+
# processor3 = Pix2StructProcessor.from_pretrained("google/pix2struct-docvqa-large")
|
138 |
|
139 |
|
140 |
+
# inputs = processor3(images=image, text=que, return_tensors="pt")
|
141 |
|
142 |
+
# predictions = model3.generate(**inputs)
|
143 |
+
# return processor3.decode(predictions[0], skip_special_tokens=True)
|
144 |
|
145 |
|
146 |
|