Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -33,15 +33,6 @@ pipe_logo.load_lora_weights("artificialguybr/LogoRedmond-LogoLoraForSDXL", weigh
|
|
33 |
pipe_logo.set_adapters(["logo"])
|
34 |
pipe_logo.to("cuda")
|
35 |
|
36 |
-
help_text = """
|
37 |
-
To optimize image results:
|
38 |
-
- Adjust the **Image CFG weight** if the image isn't changing enough or is changing too much. Lower it to allow bigger changes, or raise it to preserve original details.
|
39 |
-
- Modify the **Text CFG weight** to influence how closely the edit follows text instructions. Increase it to adhere more to the text, or decrease it for subtler changes.
|
40 |
-
- Experiment with different **random seeds** and **CFG values** for varied outcomes.
|
41 |
-
- **Rephrase your instructions** for potentially better results.
|
42 |
-
- **Increase the number of steps** for enhanced edits.
|
43 |
-
"""
|
44 |
-
|
45 |
def set_timesteps_patched(self, num_inference_steps: int, device = None):
|
46 |
self.num_inference_steps = num_inference_steps
|
47 |
|
@@ -167,7 +158,7 @@ examples=[
|
|
167 |
],
|
168 |
]
|
169 |
|
170 |
-
with gr.Blocks(css=css) as demo:
|
171 |
gr.Markdown("# Image Generator Pro")
|
172 |
with gr.Row():
|
173 |
with gr.Column(scale=4):
|
@@ -205,8 +196,7 @@ with gr.Blocks(css=css) as demo:
|
|
205 |
outputs=[input_image],
|
206 |
cache_examples=False,
|
207 |
)
|
208 |
-
|
209 |
-
gr.Markdown(help_text)
|
210 |
instruction.change(fn=response, inputs=[instruction,input_image], outputs=type, queue=False)
|
211 |
input_image.upload(fn=response, inputs=[instruction,input_image], outputs=type, queue=False)
|
212 |
|
|
|
33 |
pipe_logo.set_adapters(["logo"])
|
34 |
pipe_logo.to("cuda")
|
35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
def set_timesteps_patched(self, num_inference_steps: int, device = None):
|
37 |
self.num_inference_steps = num_inference_steps
|
38 |
|
|
|
158 |
],
|
159 |
]
|
160 |
|
161 |
+
with gr.Blocks(css=css, theme="Alyx_Theme") as demo:
|
162 |
gr.Markdown("# Image Generator Pro")
|
163 |
with gr.Row():
|
164 |
with gr.Column(scale=4):
|
|
|
196 |
outputs=[input_image],
|
197 |
cache_examples=False,
|
198 |
)
|
199 |
+
|
|
|
200 |
instruction.change(fn=response, inputs=[instruction,input_image], outputs=type, queue=False)
|
201 |
input_image.upload(fn=response, inputs=[instruction,input_image], outputs=type, queue=False)
|
202 |
|