ktrndy commited on
Commit
1168b4e
·
verified ·
1 Parent(s): 2be0274

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +42 -40
app.py CHANGED
@@ -14,6 +14,7 @@ if torch.cuda.is_available():
14
  else:
15
  torch_dtype = torch.float32
16
 
 
17
  MAX_IMAGE_SIZE = 1024
18
 
19
 
@@ -61,39 +62,55 @@ css = """
61
 
62
  with gr.Blocks(css=css) as demo:
63
  with gr.Column(elem_id="col-container"):
64
- gr.Markdown(" # Text-to-Image Gradio Template")
65
 
66
  with gr.Row():
67
- prompt = gr.Text(
68
- label="Prompt",
69
- show_label=False,
70
  max_lines=1,
71
- placeholder="Enter your prompt",
72
- container=False,
73
  )
74
 
75
- run_button = gr.Button("Run", scale=0, variant="primary")
 
 
 
 
76
 
77
- result = gr.Image(label="Result", show_label=False)
 
 
 
 
78
 
79
- with gr.Accordion("Advanced Settings", open=False):
80
- negative_prompt = gr.Text(
81
- label="Negative prompt",
82
- max_lines=1,
83
- placeholder="Enter a negative prompt",
84
- visible=False,
85
- )
86
-
87
- seed = gr.Slider(
88
  label="Seed",
89
  minimum=0,
90
  maximum=MAX_SEED,
91
  step=1,
92
- value=0,
93
  )
94
 
95
- randomize_seed = gr.Checkbox(label="Randomize seed", value=True)
 
 
 
 
 
 
 
96
 
 
 
 
 
 
 
 
 
 
97
  with gr.Row():
98
  width = gr.Slider(
99
  label="Width",
@@ -110,39 +127,24 @@ with gr.Blocks(css=css) as demo:
110
  step=32,
111
  value=1024, # Replace with defaults that work for your model
112
  )
113
-
114
- with gr.Row():
115
- guidance_scale = gr.Slider(
116
- label="Guidance scale",
117
- minimum=0.0,
118
- maximum=10.0,
119
- step=0.1,
120
- value=0.0, # Replace with defaults that work for your model
121
- )
122
-
123
- num_inference_steps = gr.Slider(
124
- label="Number of inference steps",
125
- minimum=1,
126
- maximum=50,
127
- step=1,
128
- value=2, # Replace with defaults that work for your model
129
- )
130
-
131
- gr.Examples(examples=examples, inputs=[prompt])
132
  gr.on(
133
  triggers=[run_button.click, prompt.submit],
134
  fn=infer,
135
  inputs=[
 
136
  prompt,
137
  negative_prompt,
138
  seed,
139
- randomize_seed,
140
  width,
141
  height,
142
  guidance_scale,
143
  num_inference_steps,
144
  ],
145
- outputs=[result, seed],
146
  )
147
 
148
  if __name__ == "__main__":
 
14
  else:
15
  torch_dtype = torch.float32
16
 
17
+ MAX_SEED = np.iinfo(np.int32).max
18
  MAX_IMAGE_SIZE = 1024
19
 
20
 
 
62
 
63
  with gr.Blocks(css=css) as demo:
64
  with gr.Column(elem_id="col-container"):
65
+ gr.Markdown(" # Text-to-Image demo")
66
 
67
  with gr.Row():
68
+ model_id = gr.Textbox(
69
+ label="Model ID",
 
70
  max_lines=1,
71
+ placeholder="Enter model id",
72
+ value=model_id_default,
73
  )
74
 
75
+ prompt = gr.Textbox(
76
+ label="Prompt",
77
+ max_lines=1,
78
+ placeholder="Enter your prompt",
79
+ )
80
 
81
+ negative_prompt = gr.Textbox(
82
+ label="Negative prompt",
83
+ max_lines=1,
84
+ placeholder="Enter your negative prompt",
85
+ )
86
 
87
+ with gr.Row():
88
+ seed = gr.Number(
 
 
 
 
 
 
 
89
  label="Seed",
90
  minimum=0,
91
  maximum=MAX_SEED,
92
  step=1,
93
+ value=42,
94
  )
95
 
96
+ with gr.Row():
97
+ guidance_scale = gr.Slider(
98
+ label="Guidance scale",
99
+ minimum=0.0,
100
+ maximum=10.0,
101
+ step=0.1,
102
+ value=7.0, # Replace with defaults that work for your model
103
+ )
104
 
105
+ num_inference_steps = gr.Slider(
106
+ label="Number of inference steps",
107
+ minimum=1,
108
+ maximum=50,
109
+ step=1,
110
+ value=20, # Replace with defaults that work for your model
111
+ )
112
+
113
+ with gr.Accordion("Optional Settings", open=False):
114
  with gr.Row():
115
  width = gr.Slider(
116
  label="Width",
 
127
  step=32,
128
  value=1024, # Replace with defaults that work for your model
129
  )
130
+
131
+ run_button = gr.Button("Run", scale=0, variant="primary")
132
+ result = gr.Image(label="Result", show_label=False)
133
+
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
134
  gr.on(
135
  triggers=[run_button.click, prompt.submit],
136
  fn=infer,
137
  inputs=[
138
+ model_id,
139
  prompt,
140
  negative_prompt,
141
  seed,
 
142
  width,
143
  height,
144
  guidance_scale,
145
  num_inference_steps,
146
  ],
147
+ outputs=[result],
148
  )
149
 
150
  if __name__ == "__main__":