Update my_model/tabs/run_inference.py
Browse files
my_model/tabs/run_inference.py
CHANGED
|
@@ -48,8 +48,9 @@ class InferenceRunner(StateManager):
|
|
| 48 |
"""
|
| 49 |
free_gpu_resources()
|
| 50 |
answer = model.generate_answer(question, caption, detected_objects_str)
|
|
|
|
| 51 |
free_gpu_resources()
|
| 52 |
-
return answer
|
| 53 |
|
| 54 |
|
| 55 |
def image_qa_app(self, kbvqa):
|
|
@@ -120,13 +121,13 @@ class InferenceRunner(StateManager):
|
|
| 120 |
else:
|
| 121 |
if nested_col22.button('Get Answer', key=f'answer_{image_key}', disabled=self.is_widget_disabled):
|
| 122 |
|
| 123 |
-
answer = self.answer_question(image_data['caption'], image_data['detected_objects_str'], question, kbvqa)
|
| 124 |
st.session_state['loading_in_progress'] = False
|
| 125 |
-
self.add_to_qa_history(image_key, question, answer)
|
| 126 |
|
| 127 |
-
# Display Q&A history for each image
|
| 128 |
-
for num, (q, a) in enumerate(qa_history):
|
| 129 |
-
nested_col22.text(f"Q{num+1}: {q}\nA{num+1}: {a}\n")
|
| 130 |
|
| 131 |
|
| 132 |
def run_inference(self):
|
|
|
|
| 48 |
"""
|
| 49 |
free_gpu_resources()
|
| 50 |
answer = model.generate_answer(question, caption, detected_objects_str)
|
| 51 |
+
prompt_length = modelcurrent_prompt_length
|
| 52 |
free_gpu_resources()
|
| 53 |
+
return answer, prompt_length
|
| 54 |
|
| 55 |
|
| 56 |
def image_qa_app(self, kbvqa):
|
|
|
|
| 121 |
else:
|
| 122 |
if nested_col22.button('Get Answer', key=f'answer_{image_key}', disabled=self.is_widget_disabled):
|
| 123 |
|
| 124 |
+
answer, prompt_length = self.answer_question(image_data['caption'], image_data['detected_objects_str'], question, kbvqa)
|
| 125 |
st.session_state['loading_in_progress'] = False
|
| 126 |
+
self.add_to_qa_history(image_key, question, answer, prompt_length)
|
| 127 |
|
| 128 |
+
# Display Q&A history and prompts lengths for each image
|
| 129 |
+
for num, (q, a, p) in enumerate(qa_history):
|
| 130 |
+
nested_col22.text(f"Q{num+1}: {q}\nA{num+1}: {a}\nPrompt Length: {p}\n")
|
| 131 |
|
| 132 |
|
| 133 |
def run_inference(self):
|