Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -29,7 +29,7 @@ torch.backends.cudnn.allow_tf32 = False
|
|
29 |
torch.backends.cudnn.deterministic = False
|
30 |
#torch.backends.cudnn.benchmark = False
|
31 |
torch.backends.cuda.preferred_blas_library="cublas"
|
32 |
-
torch.backends.cuda.preferred_linalg_library="cusolver"
|
33 |
|
34 |
torch.set_float32_matmul_precision("highest")
|
35 |
|
@@ -53,7 +53,7 @@ torch_dtype = torch.bfloat16
|
|
53 |
|
54 |
checkpoint = "microsoft/Phi-3.5-mini-instruct"
|
55 |
#vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16)
|
56 |
-
vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16", torch_dtype=torch.bfloat16
|
57 |
#vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16")
|
58 |
|
59 |
pipe = StableDiffusion3Pipeline.from_pretrained("ford442/stable-diffusion-3.5-medium-bf16", torch_dtype=torch.bfloat16).to(torch.device("cuda:0"))
|
|
|
29 |
torch.backends.cudnn.deterministic = False
|
30 |
#torch.backends.cudnn.benchmark = False
|
31 |
torch.backends.cuda.preferred_blas_library="cublas"
|
32 |
+
#torch.backends.cuda.preferred_linalg_library="cusolver"
|
33 |
|
34 |
torch.set_float32_matmul_precision("highest")
|
35 |
|
|
|
53 |
|
54 |
checkpoint = "microsoft/Phi-3.5-mini-instruct"
|
55 |
#vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16)
|
56 |
+
vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16", torch_dtype=torch.bfloat16)
|
57 |
#vae = AutoencoderKL.from_pretrained("ford442/sdxl-vae-bf16")
|
58 |
|
59 |
pipe = StableDiffusion3Pipeline.from_pretrained("ford442/stable-diffusion-3.5-medium-bf16", torch_dtype=torch.bfloat16).to(torch.device("cuda:0"))
|