Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -206,7 +206,7 @@ def load_and_prepare_model():
|
|
206 |
pipe.text_encoder.eval()
|
207 |
'''
|
208 |
#pipe.unet = pipe.unet.to(memory_format=torch.contiguous_format)
|
209 |
-
pipe.load_lora_weights("ford442/sdxl-vae-bf16", weight_name="LoRA/FLUX-dev-lora-add_details.safetensors", low_cpu_mem_usage=False)
|
210 |
#pipe.unet.to(memory_format=torch.channels_last)
|
211 |
#pipe.enable_vae_tiling()
|
212 |
#pipe.unet = torch.compile(pipe.unet, backend="hidet", dynamic=False, mode='max-autotune') #.to(device=device, dtype=torch.bfloat16)
|
|
|
206 |
pipe.text_encoder.eval()
|
207 |
'''
|
208 |
#pipe.unet = pipe.unet.to(memory_format=torch.contiguous_format)
|
209 |
+
#pipe.load_lora_weights("ford442/sdxl-vae-bf16", weight_name="LoRA/FLUX-dev-lora-add_details.safetensors", low_cpu_mem_usage=False)
|
210 |
#pipe.unet.to(memory_format=torch.channels_last)
|
211 |
#pipe.enable_vae_tiling()
|
212 |
#pipe.unet = torch.compile(pipe.unet, backend="hidet", dynamic=False, mode='max-autotune') #.to(device=device, dtype=torch.bfloat16)
|