Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -98,7 +98,7 @@ def load_and_prepare_model(model_id):
|
|
98 |
torch_dtype=torch.bfloat16,
|
99 |
use_safetensors=True,
|
100 |
add_watermarker=False,
|
101 |
-
)
|
102 |
pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config)
|
103 |
|
104 |
if USE_TORCH_COMPILE:
|
@@ -143,7 +143,7 @@ def generate(
|
|
143 |
):
|
144 |
global models
|
145 |
pipe = models[model_choice]
|
146 |
-
|
147 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
148 |
generator = torch.Generator(device=device).manual_seed(seed)
|
149 |
|
@@ -192,7 +192,8 @@ def generate_cpu(
|
|
192 |
):
|
193 |
global models
|
194 |
pipe = models[model_choice]
|
195 |
-
|
|
|
196 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
197 |
generator = torch.Generator(device=device).manual_seed(seed)
|
198 |
|
|
|
98 |
torch_dtype=torch.bfloat16,
|
99 |
use_safetensors=True,
|
100 |
add_watermarker=False,
|
101 |
+
)
|
102 |
pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config)
|
103 |
|
104 |
if USE_TORCH_COMPILE:
|
|
|
143 |
):
|
144 |
global models
|
145 |
pipe = models[model_choice]
|
146 |
+
pipe.to("cuda:0")
|
147 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
148 |
generator = torch.Generator(device=device).manual_seed(seed)
|
149 |
|
|
|
192 |
):
|
193 |
global models
|
194 |
pipe = models[model_choice]
|
195 |
+
pipe.to("cpu")
|
196 |
+
|
197 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
198 |
generator = torch.Generator(device=device).manual_seed(seed)
|
199 |
|