prithivMLmods commited on
Commit
302b5fb
·
verified ·
1 Parent(s): adb46f7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -217,9 +217,9 @@ def adjust_generation_mode(speed_mode):
217
  if speed_mode == "Speed (8 steps)":
218
  return gr.update(value="Speed mode selected - 8 steps with Lightning LoRA"), 8, 1.0
219
  else:
220
- return gr.update(value="Quality mode selected - 45 steps for best quality"), 45, 3.5
221
 
222
- @spaces.GPU(duration=70)
223
  def create_image(prompt_mash, steps, seed, cfg_scale, width, height, lora_scale, negative_prompt=""):
224
  pipe.to("cuda")
225
  generator = torch.Generator(device="cuda").manual_seed(seed)
@@ -238,7 +238,7 @@ def create_image(prompt_mash, steps, seed, cfg_scale, width, height, lora_scale,
238
 
239
  return image
240
 
241
- @spaces.GPU(duration=70)
242
  def process_adapter_generation(prompt, cfg_scale, steps, selected_index, randomize_seed, seed, aspect_ratio, lora_scale, speed_mode, progress=gr.Progress(track_tqdm=True)):
243
  if selected_index is None:
244
  raise gr.Error("You must select a LoRA before proceeding.")
 
217
  if speed_mode == "Speed (8 steps)":
218
  return gr.update(value="Speed mode selected - 8 steps with Lightning LoRA"), 8, 1.0
219
  else:
220
+ return gr.update(value="Quality mode selected - 48 steps for best quality"), 48, 3.5
221
 
222
+ @spaces.GPU(duration=108)
223
  def create_image(prompt_mash, steps, seed, cfg_scale, width, height, lora_scale, negative_prompt=""):
224
  pipe.to("cuda")
225
  generator = torch.Generator(device="cuda").manual_seed(seed)
 
238
 
239
  return image
240
 
241
+ @spaces.GPU(duration=108)
242
  def process_adapter_generation(prompt, cfg_scale, steps, selected_index, randomize_seed, seed, aspect_ratio, lora_scale, speed_mode, progress=gr.Progress(track_tqdm=True)):
243
  if selected_index is None:
244
  raise gr.Error("You must select a LoRA before proceeding.")