Junjie96 commited on
Commit
90fc7e5
·
verified ·
1 Parent(s): 889d8af

Upload 45 files

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. app.py +5 -37
  3. requirements.txt +1 -0
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  title: AnyStory
3
- emoji: 💻
4
  colorFrom: indigo
5
  colorTo: yellow
6
  sdk: gradio
 
1
  ---
2
  title: AnyStory
3
+ emoji: ✏️
4
  colorFrom: indigo
5
  colorTo: yellow
6
  sdk: gradio
app.py CHANGED
@@ -10,7 +10,6 @@ import gradio as gr
10
  import numpy as np
11
  from PIL import Image
12
 
13
- from src.agent import extend_prompt, translate_prompt
14
  from src.anystory import call_anystory
15
  from src.matting import ImageUniversalMatting
16
  from src.util import upload_pil_2_oss
@@ -112,8 +111,6 @@ def interface():
112
 
113
  with gr.Group():
114
  prompt = gr.Textbox(value="", label='Prompt', lines=6, show_label=True)
115
- en_prompt = gr.Textbox(value="", label='prompt', lines=6, show_label=True, visible=False)
116
- # prompt_extend_button = gr.Button(value="提示词扩写")
117
 
118
  with gr.Column(scale=1, min_width=100):
119
  result_gallery = gr.Image(type="pil", label="Generated Image", visible=True, height=450)
@@ -139,23 +136,11 @@ def interface():
139
  fn=set_image_seg_finished, outputs=generated_information
140
  )
141
 
142
- # prompt_extend_button.click(
143
- # fn=set_prompt_extend_unfinished, outputs=generated_information
144
- # ).then(
145
- # fn=extend_prompt, inputs=[prompt], outputs=[prompt]
146
- # ).then(
147
- # fn=set_prompt_extend_finished, outputs=generated_information
148
- # )
149
-
150
  run_button.click(
151
- fn=set_prompt_translate_unfinished, outputs=generated_information
152
- ).then(
153
- fn=translate_prompt, inputs=[prompt], outputs=[en_prompt]
154
- ).then(
155
  fn=set_image_generate_unfinished, outputs=generated_information
156
  ).then(
157
  fn=process,
158
- inputs=[pil_subject_A_image, pil_subject_A_mask, pil_subject_B_image, pil_subject_B_mask, en_prompt],
159
  outputs=[result_gallery]
160
  ).then(
161
  fn=set_image_generate_finished, outputs=generated_information
@@ -247,7 +232,8 @@ def interface():
247
  gr.Examples(
248
  label="Examples",
249
  examples=examples,
250
- inputs=[pil_subject_A_image, pil_subject_A_mask, pil_subject_B_image, pil_subject_B_mask, prompt, result_gallery],
 
251
  )
252
 
253
 
@@ -262,24 +248,6 @@ def set_image_seg_finished():
262
  return gr.update(visible=True, value="<h3>Subject mask ready!</h3>")
263
 
264
 
265
- def set_prompt_extend_unfinished():
266
- return gr.update(
267
- visible=True,
268
- value="<h3>(Unfinished) Rewriting your prompt... ✍️</h3>",
269
- )
270
-
271
-
272
- def set_prompt_extend_finished():
273
- return gr.update(visible=True, value="<h3>Prompt expanded successfully!</h3>")
274
-
275
-
276
- def set_prompt_translate_unfinished():
277
- return gr.update(
278
- visible=True,
279
- value="<h3>(Unfinished) Preprocessing...</h3>",
280
- )
281
-
282
-
283
  def set_image_generate_unfinished():
284
  return gr.update(
285
  visible=True,
@@ -301,7 +269,7 @@ if __name__ == "__main__":
301
  &nbsp;
302
  <a href='https://aigcdesigngroup.github.io/AnyStory/'><img src='https://img.shields.io/badge/Project_Page-AnyStory-green' alt='Project Page'></a>
303
  &nbsp;
304
- <a href='https://modelscope.cn/studios/damo/studio_anytext'><img src='https://img.shields.io/badge/Demo-ModelScope-blue'></a>
305
  </div>
306
  </br>
307
  </div>
@@ -313,7 +281,7 @@ if __name__ == "__main__":
313
  """
314
 
315
  description = r"""🚀🚀🚀 Quick Start:<br>
316
- 1. Upload subject reference images (clean background; real human IDs unsupported for now), Add prompts (CN/EN supported), and Click "<b>RUN</b>".<br>
317
  2. (Recommended) Click "<b>Segment Subject</b>" to create masks (or upload your own B&W masks) for subjects. This helps the model better reference the subject you specify (otherwise, we will perform automatic detection). 🤗<br>
318
  """
319
 
 
10
  import numpy as np
11
  from PIL import Image
12
 
 
13
  from src.anystory import call_anystory
14
  from src.matting import ImageUniversalMatting
15
  from src.util import upload_pil_2_oss
 
111
 
112
  with gr.Group():
113
  prompt = gr.Textbox(value="", label='Prompt', lines=6, show_label=True)
 
 
114
 
115
  with gr.Column(scale=1, min_width=100):
116
  result_gallery = gr.Image(type="pil", label="Generated Image", visible=True, height=450)
 
136
  fn=set_image_seg_finished, outputs=generated_information
137
  )
138
 
 
 
 
 
 
 
 
 
139
  run_button.click(
 
 
 
 
140
  fn=set_image_generate_unfinished, outputs=generated_information
141
  ).then(
142
  fn=process,
143
+ inputs=[pil_subject_A_image, pil_subject_A_mask, pil_subject_B_image, pil_subject_B_mask, prompt],
144
  outputs=[result_gallery]
145
  ).then(
146
  fn=set_image_generate_finished, outputs=generated_information
 
232
  gr.Examples(
233
  label="Examples",
234
  examples=examples,
235
+ inputs=[pil_subject_A_image, pil_subject_A_mask, pil_subject_B_image, pil_subject_B_mask, prompt,
236
+ result_gallery],
237
  )
238
 
239
 
 
248
  return gr.update(visible=True, value="<h3>Subject mask ready!</h3>")
249
 
250
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
251
  def set_image_generate_unfinished():
252
  return gr.update(
253
  visible=True,
 
269
  &nbsp;
270
  <a href='https://aigcdesigngroup.github.io/AnyStory/'><img src='https://img.shields.io/badge/Project_Page-AnyStory-green' alt='Project Page'></a>
271
  &nbsp;
272
+ <a href='https://modelscope.cn/studios/iic/AnyStory'><img src='https://img.shields.io/badge/Demo-ModelScope-blue'></a>
273
  </div>
274
  </br>
275
  </div>
 
281
  """
282
 
283
  description = r"""🚀🚀🚀 Quick Start:<br>
284
+ 1. Upload subject reference images (clean background; real human IDs unsupported for now), Add prompts (only EN supported), and Click "<b>RUN</b>".<br>
285
  2. (Recommended) Click "<b>Segment Subject</b>" to create masks (or upload your own B&W masks) for subjects. This helps the model better reference the subject you specify (otherwise, we will perform automatic detection). 🤗<br>
286
  """
287
 
requirements.txt CHANGED
@@ -2,3 +2,4 @@ oss2
2
  phidata
3
  dashscope
4
  tensorflow==2.15
 
 
2
  phidata
3
  dashscope
4
  tensorflow==2.15
5
+ openai