ford442 commited on
Commit
0daa48f
·
verified ·
1 Parent(s): 81eb9fa

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -3
app.py CHANGED
@@ -109,7 +109,8 @@ def apply_style(style_name: str, positive: str, negative: str = "") -> Tuple[str
109
 
110
  def load_and_prepare_model():
111
  #vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16", safety_checker=None)
112
- vaeX = AutoencoderKL.from_pretrained("stabilityai/sdxl-vae", safety_checker=None,use_safetensors=False,token=HF_TOKEN)
 
113
  #vae = AutoencoderKL.from_single_file('https://huggingface.co/ford442/sdxl-vae-bf16/mySLR/myslrVAE_v10.safetensors')
114
  #vaeX = AutoencoderKL.from_pretrained('ford442/Juggernaut-XI-v11-fp32',subfolder='vae') # ,use_safetensors=True FAILS
115
  #unetX = UNet2DConditionModel.from_pretrained('ford442/RealVisXL_V5.0_BF16',subfolder='unet').to(torch.bfloat16) # ,use_safetensors=True FAILS
@@ -169,7 +170,7 @@ def load_and_prepare_model():
169
  #pipe.scheduler = DPMSolverMultistepScheduler.from_pretrained('SG161222/RealVisXL_V5.0', subfolder='scheduler', algorithm_type='sde-dpmsolver++')
170
 
171
  #**** NEVER TO BFLOAT BEFORE CUDA****#
172
- pipe.vae = vaeX #.to(torch.bfloat16)
173
  #**** NEVER TO BFLOAT BEFORE CUDA****#
174
 
175
  #pipe.unet = unetX
@@ -187,7 +188,7 @@ def load_and_prepare_model():
187
  # pipe.load_lora_weights("ford442/sdxl-vae-bf16", weight_name="LoRA/Fantasy_World_XL.safetensors", adapter_name="fantasy")
188
  pipe.vae.set_default_attn_processor() # Set attention processor first
189
 
190
- pipe.load_lora_weights("Keltezaa/flux_pussy_NSFW", weight_name="flux-pussy.safetensors", adapter_name="vagina", low_cpu_mem_usage=False,token=HF_TOKEN)
191
  pipe.load_lora_weights("xey/sldr_flux_nsfw_v2-studio", weight_name="sldr_flux_nsfw_v2-studio.safetensors", adapter_name="nsfw", low_cpu_mem_usage=False,token=HF_TOKEN)
192
  pipe.load_lora_weights("ford442/sdxl-vae-bf16", weight_name="LoRA/skin_texture_style_v4.safetensors", adapter_name="skin", low_cpu_mem_usage=False,token=HF_TOKEN)
193
  #pipe.unet.load_lora_adapter("ford442/sdxl-vae-bf16", weight_name="LoRA/skin_texture_style_v4.safetensors", prefix="unet")
 
109
 
110
  def load_and_prepare_model():
111
  #vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16", safety_checker=None)
112
+ #vaeX = AutoencoderKL.from_pretrained("stabilityai/sdxl-vae", safety_checker=None,use_safetensors=False,token=HF_TOKEN)
113
+ vaeRV = AutoencoderKL.from_pretrained("SG161222/RealVisXL_V5.0", subfolder='vae', safety_checker=None,token=HF_TOKEN)
114
  #vae = AutoencoderKL.from_single_file('https://huggingface.co/ford442/sdxl-vae-bf16/mySLR/myslrVAE_v10.safetensors')
115
  #vaeX = AutoencoderKL.from_pretrained('ford442/Juggernaut-XI-v11-fp32',subfolder='vae') # ,use_safetensors=True FAILS
116
  #unetX = UNet2DConditionModel.from_pretrained('ford442/RealVisXL_V5.0_BF16',subfolder='unet').to(torch.bfloat16) # ,use_safetensors=True FAILS
 
170
  #pipe.scheduler = DPMSolverMultistepScheduler.from_pretrained('SG161222/RealVisXL_V5.0', subfolder='scheduler', algorithm_type='sde-dpmsolver++')
171
 
172
  #**** NEVER TO BFLOAT BEFORE CUDA****#
173
+ pipe.vae = vaeRV #.to(torch.bfloat16)
174
  #**** NEVER TO BFLOAT BEFORE CUDA****#
175
 
176
  #pipe.unet = unetX
 
188
  # pipe.load_lora_weights("ford442/sdxl-vae-bf16", weight_name="LoRA/Fantasy_World_XL.safetensors", adapter_name="fantasy")
189
  pipe.vae.set_default_attn_processor() # Set attention processor first
190
 
191
+ #pipe.load_lora_weights("Keltezaa/flux_pussy_NSFW", weight_name="flux-pussy.safetensors", adapter_name="vagina", low_cpu_mem_usage=False,token=HF_TOKEN)
192
  pipe.load_lora_weights("xey/sldr_flux_nsfw_v2-studio", weight_name="sldr_flux_nsfw_v2-studio.safetensors", adapter_name="nsfw", low_cpu_mem_usage=False,token=HF_TOKEN)
193
  pipe.load_lora_weights("ford442/sdxl-vae-bf16", weight_name="LoRA/skin_texture_style_v4.safetensors", adapter_name="skin", low_cpu_mem_usage=False,token=HF_TOKEN)
194
  #pipe.unet.load_lora_adapter("ford442/sdxl-vae-bf16", weight_name="LoRA/skin_texture_style_v4.safetensors", prefix="unet")