3v324v23 commited on
Commit
d21fbe9
·
1 Parent(s): 0b42636

preprocess reopen

Browse files
Files changed (1) hide show
  1. app.py +21 -8
app.py CHANGED
@@ -60,14 +60,14 @@ controlnet_path = OrderedDict([
60
 
61
  preprocess_method = [
62
  'canny' ,
63
- # 'depth' ,
64
  # 'hed' ,
65
- # 'mlsd' ,
66
  # 'normal' ,
67
- # 'openpose' ,
68
- # 'openpose_withface' ,
69
- # 'openpose_withfacehand',
70
- # 'scribble' ,
71
  'none' ,
72
  ]
73
 
@@ -414,9 +414,9 @@ def interface():
414
  button = gr.Button("Run")
415
  with gr.Column():
416
  ctl_input = gr.Image(label='Control Input', type='pil', elem_id='customized_imbox')
417
- do_preprocess = gr.Checkbox(label='Preprocess (Disabled)', value=False)
418
  with gr.Row():
419
- ctl_method = gr.Dropdown(label='Preprocess Type (Fixed to none)', choices=preprocess_method, value='canny')
420
  tag_ctl = gr.Dropdown(label='ControlNet', choices=[pi for pi in controlnet_path.keys()], value='canny')
421
  with gr.Column():
422
  img_output = gr.Gallery(label="Image Result", elem_id='customized_imbox').style(grid=n_sample_image+1)
@@ -513,6 +513,19 @@ if True:
513
  <h1 style="font-weight: 900; font-size: 3rem; margin: 0rem">
514
  Prompt-Free Diffusion
515
  </h1>
 
 
 
 
 
 
 
 
 
 
 
 
 
516
  </div>
517
  """)
518
 
 
60
 
61
  preprocess_method = [
62
  'canny' ,
63
+ 'depth' ,
64
  # 'hed' ,
65
+ 'mlsd' ,
66
  # 'normal' ,
67
+ 'openpose' ,
68
+ 'openpose_withface' ,
69
+ 'openpose_withfacehand',
70
+ 'scribble' ,
71
  'none' ,
72
  ]
73
 
 
414
  button = gr.Button("Run")
415
  with gr.Column():
416
  ctl_input = gr.Image(label='Control Input', type='pil', elem_id='customized_imbox')
417
+ do_preprocess = gr.Checkbox(label='Preprocess', value=False)
418
  with gr.Row():
419
+ ctl_method = gr.Dropdown(label='Preprocess Type', choices=preprocess_method, value='canny')
420
  tag_ctl = gr.Dropdown(label='ControlNet', choices=[pi for pi in controlnet_path.keys()], value='canny')
421
  with gr.Column():
422
  img_output = gr.Gallery(label="Image Result", elem_id='customized_imbox').style(grid=n_sample_image+1)
 
513
  <h1 style="font-weight: 900; font-size: 3rem; margin: 0rem">
514
  Prompt-Free Diffusion
515
  </h1>
516
+ <p style="font-size: 1rem; margin: 0rem">
517
+ Xingqian Xu<sup>1,6</sup>, Jiayi Guo<sup>1,2</sup>, Zhangyang Wang<sup>3,6</sup>, Gao Huang<sup>2</sup>, Irfan Essa<sup>4,5</sup>, and Humphrey Shi<sup>1,6</sup>
518
+ </p>
519
+ <p style="font-size: 0.8rem; margin: 0rem; line-height: 1em">
520
+ <sup>1</sup>SHI Labs @ UIUC & Oregon, <sup>2</sup>Tsinghua University, <sup>3</sup>UT Austin, <sup>4</sup>Georgia Tech, <sup>5</sup>Google Research, <sup>6</sup>Picsart AI Research (PAIR)
521
+ </p>
522
+ <p style="font-size: 0.9rem; margin: 0rem; line-height: 1.2em; margin-top:1em">
523
+ The performance of Text2Image is largely dependent on text prompts.
524
+ In Prompt-Free Diffusion, no prompt is needed, just a reference images!
525
+ At the core of Prompt-Free Diffusion is an image-only semantic context encoder (SeeCoder).
526
+ SeeCoder is reusable to any customized T2I models: just drop in and replace CLIP, then you will create your own prompt-free diffusion.
527
+ <a href="https://github.com/SHI-Labs/Prompt-Free-Diffusion">[Github]</a> <a href="https://arxiv.org/abs/2305.16223">[arXiv]</a>
528
+ </p>
529
  </div>
530
  """)
531