Update app.py
Browse files
app.py
CHANGED
@@ -138,16 +138,24 @@ def predict(_chatbot, task_history) -> tuple:
|
|
138 |
tokenized_inputs = tokenizer(inputs, return_tensors='pt')
|
139 |
tokenized_inputs = tokenized_inputs.to(model.device)
|
140 |
pred = model.generate(**tokenized_inputs)
|
141 |
-
response = tokenizer.decode(pred.cpu()[0], skip_special_tokens=False)
|
142 |
if 'image' in query[0]:
|
143 |
print("Model response:", response)
|
144 |
-
|
145 |
-
|
146 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
147 |
else:
|
148 |
-
|
149 |
-
_chatbot[-1] = (chat_query, text_response)
|
150 |
-
task_history[-1] = (chat_query, text_response)
|
151 |
|
152 |
return _chatbot, task_history
|
153 |
|
|
|
138 |
tokenized_inputs = tokenizer(inputs, return_tensors='pt')
|
139 |
tokenized_inputs = tokenized_inputs.to(model.device)
|
140 |
pred = model.generate(**tokenized_inputs)
|
141 |
+
response = tokenizer.decode(pred.cpu()[0], skip_special_tokens=False)
|
142 |
if 'image' in query[0]:
|
143 |
print("Model response:", response)
|
144 |
+
image = tokenizer.draw_bbox_on_latest_picture(response)
|
145 |
+
if image is not None:
|
146 |
+
image_path = save_image(image, uploaded_file_dir)
|
147 |
+
formatted_response = (chat_query, image_path)
|
148 |
+
else:
|
149 |
+
formatted_response = (chat_query, response)
|
150 |
+
else:
|
151 |
+
text_response = response.strip()
|
152 |
+
formatted_response = (chat_query, text_response)
|
153 |
+
|
154 |
+
_chatbot[-1] = formatted_response
|
155 |
+
if task_history:
|
156 |
+
task_history[-1] = formatted_response
|
157 |
else:
|
158 |
+
task_history.append(formatted_response)
|
|
|
|
|
159 |
|
160 |
return _chatbot, task_history
|
161 |
|