Manjushri commited on
Commit
2cf05e4
·
verified ·
1 Parent(s): d152942

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -4
app.py CHANGED
@@ -44,12 +44,18 @@ def genie (prompt, negative_prompt, height, width, scale, steps, seed, upscaling
44
  torch.cuda.empty_cache()
45
  return (image, upscaled)
46
  else:
47
- image = refiner(prompt=prompt, prompt_2=prompt_2, negative_prompt=negative_prompt, negative_prompt_2=negative_prompt_2, image=int_image, num_inference_steps=n_steps ,denoising_start=high_noise_frac).images[0]
48
- torch.cuda.empty_cache()
49
- return (image, image)
50
  else:
51
- image = pipe(prompt, prompt_2=prompt_2, negative_prompt=negative_prompt, negative_prompt_2=negative_prompt_2, num_inference_steps=steps, height=height, width=width, guidance_scale=scale, num_images_per_prompt=1, generator=generator).images[0]
 
 
52
  torch.cuda.empty_cache()
 
 
 
 
53
  return (image, image)
54
 
55
  gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit. A Token is Any Word, Number, Symbol, or Punctuation. Everything Over 77 Will Be Truncated!'),
 
44
  torch.cuda.empty_cache()
45
  return (image, upscaled)
46
  else:
47
+ image = refiner(prompt=prompt, prompt_2=prompt_2, negative_prompt=negative_prompt, negative_prompt_2=negative_prompt_2, image=int_image, num_inference_steps=n_steps ,denoising_start=high_noise_frac).images[0]
48
+ torch.cuda.empty_cache()
49
+ return (image, image)
50
  else:
51
+ if upscaling == 'Yes':
52
+ image = refiner(prompt=prompt, prompt_2=prompt_2, negative_prompt=negative_prompt, negative_prompt_2=negative_prompt_2, image=int_image, num_inference_steps=n_steps, denoising_start=high_noise_frac).images[0] #num_inference_steps=n_steps,
53
+ upscaled = upscaler(prompt=prompt, negative_prompt=negative_prompt, image=image, num_inference_steps=15, guidance_scale=0).images[0]
54
  torch.cuda.empty_cache()
55
+ return (image, upscaled)
56
+ else:
57
+ image = pipe(prompt, prompt_2=prompt_2, negative_prompt=negative_prompt, negative_prompt_2=negative_prompt_2, num_inference_steps=steps, height=height, width=width, guidance_scale=scale, num_images_per_prompt=1, generator=generator).images[0]
58
+ torch.cuda.empty_cache()
59
  return (image, image)
60
 
61
  gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit. A Token is Any Word, Number, Symbol, or Punctuation. Everything Over 77 Will Be Truncated!'),