Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -98,7 +98,7 @@ def load_and_prepare_model(model_id):
|
|
| 98 |
torch_dtype=torch.bfloat16,
|
| 99 |
use_safetensors=True,
|
| 100 |
add_watermarker=False,
|
| 101 |
-
)
|
| 102 |
pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config)
|
| 103 |
|
| 104 |
if USE_TORCH_COMPILE:
|
|
@@ -143,7 +143,7 @@ def generate(
|
|
| 143 |
):
|
| 144 |
global models
|
| 145 |
pipe = models[model_choice]
|
| 146 |
-
|
| 147 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
| 148 |
generator = torch.Generator(device=device).manual_seed(seed)
|
| 149 |
|
|
@@ -192,7 +192,8 @@ def generate_cpu(
|
|
| 192 |
):
|
| 193 |
global models
|
| 194 |
pipe = models[model_choice]
|
| 195 |
-
|
|
|
|
| 196 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
| 197 |
generator = torch.Generator(device=device).manual_seed(seed)
|
| 198 |
|
|
|
|
| 98 |
torch_dtype=torch.bfloat16,
|
| 99 |
use_safetensors=True,
|
| 100 |
add_watermarker=False,
|
| 101 |
+
)
|
| 102 |
pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config)
|
| 103 |
|
| 104 |
if USE_TORCH_COMPILE:
|
|
|
|
| 143 |
):
|
| 144 |
global models
|
| 145 |
pipe = models[model_choice]
|
| 146 |
+
pipe.to("cuda:0")
|
| 147 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
| 148 |
generator = torch.Generator(device=device).manual_seed(seed)
|
| 149 |
|
|
|
|
| 192 |
):
|
| 193 |
global models
|
| 194 |
pipe = models[model_choice]
|
| 195 |
+
pipe.to("cpu")
|
| 196 |
+
|
| 197 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
| 198 |
generator = torch.Generator(device=device).manual_seed(seed)
|
| 199 |
|