new
Browse files
app.py
CHANGED
@@ -17,12 +17,8 @@ from huggingface_hub import hf_hub_download
|
|
17 |
|
18 |
#gradio.helpers.CACHED_FOLDER = '/data/cache'
|
19 |
|
20 |
-
# Load the pipeline with authentication token
|
21 |
pipe = StableVideoDiffusionPipeline.from_pretrained(
|
22 |
-
"
|
23 |
-
torch_dtype=torch.float16,
|
24 |
-
variant="fp16",
|
25 |
-
use_auth_token=os.getenv("HUGGINGFACE_TOKEN") # Fetch the token from the environment variable
|
26 |
)
|
27 |
pipe.to("cuda")
|
28 |
#pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
@@ -47,7 +43,7 @@ def sample(
|
|
47 |
if image.mode == "RGBA":
|
48 |
image = image.convert("RGB")
|
49 |
|
50 |
-
if
|
51 |
seed = random.randint(0, max_64_bit_int)
|
52 |
generator = torch.manual_seed(seed)
|
53 |
|
|
|
17 |
|
18 |
#gradio.helpers.CACHED_FOLDER = '/data/cache'
|
19 |
|
|
|
20 |
pipe = StableVideoDiffusionPipeline.from_pretrained(
|
21 |
+
"stabilityai/stable-video-diffusion-img2vid-xt", torch_dtype=torch.float16, variant="fp16"
|
|
|
|
|
|
|
22 |
)
|
23 |
pipe.to("cuda")
|
24 |
#pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
|
|
43 |
if image.mode == "RGBA":
|
44 |
image = image.convert("RGB")
|
45 |
|
46 |
+
if(randomize_seed):
|
47 |
seed = random.randint(0, max_64_bit_int)
|
48 |
generator = torch.manual_seed(seed)
|
49 |
|