Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -274,8 +274,9 @@ def generate_30(
|
|
274 |
randomize_seed: bool = False,
|
275 |
use_resolution_binning: bool = True,
|
276 |
num_images: int = 1,
|
277 |
-
|
278 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
|
|
279 |
):
|
280 |
torch.backends.cudnn.benchmark = False
|
281 |
torch.cuda.empty_cache()
|
@@ -341,7 +342,7 @@ def generate_60(
|
|
341 |
randomize_seed: bool = False,
|
342 |
use_resolution_binning: bool = True,
|
343 |
num_images: int = 1,
|
344 |
-
|
345 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
346 |
):
|
347 |
torch.backends.cudnn.benchmark = True
|
@@ -408,7 +409,7 @@ def generate_90(
|
|
408 |
randomize_seed: bool = False,
|
409 |
use_resolution_binning: bool = True,
|
410 |
num_images: int = 1,
|
411 |
-
|
412 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
413 |
):
|
414 |
torch.backends.cudnn.benchmark = True
|
@@ -614,7 +615,7 @@ with gr.Blocks(theme=gr.themes.Origin(),css=css) as demo:
|
|
614 |
num_inference_steps,
|
615 |
randomize_seed,
|
616 |
num_images,
|
617 |
-
denoise
|
618 |
],
|
619 |
outputs=[result, seed],
|
620 |
)
|
@@ -638,7 +639,7 @@ with gr.Blocks(theme=gr.themes.Origin(),css=css) as demo:
|
|
638 |
num_inference_steps,
|
639 |
randomize_seed,
|
640 |
num_images,
|
641 |
-
denoise
|
642 |
],
|
643 |
outputs=[result, seed],
|
644 |
)
|
@@ -661,8 +662,8 @@ with gr.Blocks(theme=gr.themes.Origin(),css=css) as demo:
|
|
661 |
guidance_scale,
|
662 |
num_inference_steps,
|
663 |
randomize_seed,
|
664 |
-
num_images,
|
665 |
-
denoise
|
666 |
],
|
667 |
outputs=[result, seed],
|
668 |
)
|
|
|
274 |
randomize_seed: bool = False,
|
275 |
use_resolution_binning: bool = True,
|
276 |
num_images: int = 1,
|
277 |
+
denoise: float = 0.3,
|
278 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
279 |
+
|
280 |
):
|
281 |
torch.backends.cudnn.benchmark = False
|
282 |
torch.cuda.empty_cache()
|
|
|
342 |
randomize_seed: bool = False,
|
343 |
use_resolution_binning: bool = True,
|
344 |
num_images: int = 1,
|
345 |
+
denoise: float = 0.3,
|
346 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
347 |
):
|
348 |
torch.backends.cudnn.benchmark = True
|
|
|
409 |
randomize_seed: bool = False,
|
410 |
use_resolution_binning: bool = True,
|
411 |
num_images: int = 1,
|
412 |
+
denoise: float = 0.3,
|
413 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
414 |
):
|
415 |
torch.backends.cudnn.benchmark = True
|
|
|
615 |
num_inference_steps,
|
616 |
randomize_seed,
|
617 |
num_images,
|
618 |
+
denoise,
|
619 |
],
|
620 |
outputs=[result, seed],
|
621 |
)
|
|
|
639 |
num_inference_steps,
|
640 |
randomize_seed,
|
641 |
num_images,
|
642 |
+
denoise,
|
643 |
],
|
644 |
outputs=[result, seed],
|
645 |
)
|
|
|
662 |
guidance_scale,
|
663 |
num_inference_steps,
|
664 |
randomize_seed,
|
665 |
+
num_images,
|
666 |
+
denoise,
|
667 |
],
|
668 |
outputs=[result, seed],
|
669 |
)
|