ivelin commited on
Commit
f85a58b
·
1 Parent(s): dc7c41f

add model checkpoint revision as an option

Browse files
Files changed (1) hide show
  1. app.py +4 -3
app.py CHANGED
@@ -86,9 +86,10 @@ def translate_point_coords_from_out_to_in(point=None, input_image_size=None, out
86
  f"translated point={point}, resized_image_size: {resized_width, resized_height}")
87
 
88
 
89
- def process_refexp(image: Image, prompt: str):
90
 
91
  print(f"(image, prompt): {image}, {prompt}")
 
92
 
93
  # trim prompt to 80 characters and normalize to lowercase
94
  prompt = prompt[:80].lower()
@@ -184,7 +185,7 @@ def process_refexp(image: Image, prompt: str):
184
 
185
 
186
  title = "Demo: Donut 🍩 for UI RefExp - Center Point (by GuardianUI)"
187
- description = "Gradio Demo for Donut RefExp task, an instance of `VisionEncoderDecoderModel` fine-tuned on [UIBert RefExp](https://huggingface.co/datasets/ivelin/ui_refexp_saved) Dataset (UI Referring Expression). To use it, simply upload your image and type a prompt and click 'submit', or click one of the examples to load them. See the model training <a href='https://colab.research.google.com/github/ivelin/donut_ui_refexp/blob/main/Fine_tune_Donut_on_UI_RefExp.ipynb' target='_parent'>Colab Notebook</a> for this space. Read more at the links below."
188
  article = "<p style='text-align: center'><a href='https://arxiv.org/abs/2111.15664' target='_blank'>Donut: OCR-free Document Understanding Transformer</a> | <a href='https://github.com/clovaai/donut' target='_blank'>Github Repo</a></p>"
189
  examples = [["example_1.jpg", "select the setting icon from top right corner"],
190
  ["example_1.jpg", "click on down arrow beside the entertainment"],
@@ -211,7 +212,7 @@ examples = [["example_1.jpg", "select the setting icon from top right corner"],
211
  ]
212
 
213
  demo = gr.Interface(fn=process_refexp,
214
- inputs=[gr.Image(type="pil"), "text"],
215
  outputs=[gr.Image(type="pil"), "json"],
216
  title=title,
217
  description=description,
 
86
  f"translated point={point}, resized_image_size: {resized_width, resized_height}")
87
 
88
 
89
+ def process_refexp(image: Image, prompt: str, revision: str = 'main'):
90
 
91
  print(f"(image, prompt): {image}, {prompt}")
92
+ print(f"model checkpoint revision: {revision}")
93
 
94
  # trim prompt to 80 characters and normalize to lowercase
95
  prompt = prompt[:80].lower()
 
185
 
186
 
187
  title = "Demo: Donut 🍩 for UI RefExp - Center Point (by GuardianUI)"
188
+ description = "Gradio Demo for Donut RefExp task, an instance of `VisionEncoderDecoderModel` fine-tuned on [UIBert RefExp](https://huggingface.co/datasets/ivelin/ui_refexp_saved) Dataset (UI Referring Expression). To use it, simply upload your image and type a prompt and click 'submit', or click one of the examples to load them. Optionally enter value for model git revision; latest checkpoint will be used by default."
189
  article = "<p style='text-align: center'><a href='https://arxiv.org/abs/2111.15664' target='_blank'>Donut: OCR-free Document Understanding Transformer</a> | <a href='https://github.com/clovaai/donut' target='_blank'>Github Repo</a></p>"
190
  examples = [["example_1.jpg", "select the setting icon from top right corner"],
191
  ["example_1.jpg", "click on down arrow beside the entertainment"],
 
212
  ]
213
 
214
  demo = gr.Interface(fn=process_refexp,
215
+ inputs=[gr.Image(type="pil"), "text", "text"],
216
  outputs=[gr.Image(type="pil"), "json"],
217
  title=title,
218
  description=description,