Update handler.py
Browse files- handler.py +14 -13
handler.py
CHANGED
@@ -110,6 +110,12 @@ class EndpointHandler:
|
|
110 |
# self.image_to_video.text_encoder_2 = self.image_to_video.text_encoder_2.half()
|
111 |
# self.image_to_video.transformer = self.image_to_video.transformer.to(torch.bfloat16)
|
112 |
# self.image_to_video.vae = self.image_to_video.vae.half()
|
|
|
|
|
|
|
|
|
|
|
|
|
113 |
else:
|
114 |
# Initialize text-to-video pipeline
|
115 |
self.text_to_video = HunyuanVideoPipeline.from_pretrained(
|
@@ -124,6 +130,14 @@ class EndpointHandler:
|
|
124 |
self.text_to_video.transformer = self.text_to_video.transformer.to(torch.bfloat16)
|
125 |
self.text_to_video.vae = self.text_to_video.vae.half()
|
126 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
127 |
# enable FasterCache
|
128 |
|
129 |
# those values are coming from here:
|
@@ -336,21 +350,8 @@ class EndpointHandler:
|
|
336 |
config.input_image_quality,
|
337 |
)
|
338 |
generation_kwargs["image"] = processed_image
|
339 |
-
|
340 |
-
apply_enhance_a_video(self.image_to_video.transformer, EnhanceAVideoConfig(
|
341 |
-
weight=config.enhance_a_video_weight if config.enable_enhance_a_video else 0.0,
|
342 |
-
num_frames_callback=lambda: (config.num_frames - 1),
|
343 |
-
_attention_type=1
|
344 |
-
))
|
345 |
-
|
346 |
frames = self.image_to_video(**generation_kwargs).frames
|
347 |
else:
|
348 |
-
apply_enhance_a_video(self.text_to_video.transformer, EnhanceAVideoConfig(
|
349 |
-
weight=config.enhance_a_video_weight if config.enable_enhance_a_video else 0.0,
|
350 |
-
num_frames_callback=lambda: (config.num_frames - 1),
|
351 |
-
_attention_type=1
|
352 |
-
))
|
353 |
-
|
354 |
frames = self.text_to_video(**generation_kwargs).frames
|
355 |
|
356 |
|
|
|
110 |
# self.image_to_video.text_encoder_2 = self.image_to_video.text_encoder_2.half()
|
111 |
# self.image_to_video.transformer = self.image_to_video.transformer.to(torch.bfloat16)
|
112 |
# self.image_to_video.vae = self.image_to_video.vae.half()
|
113 |
+
|
114 |
+
# apply_enhance_a_video(self.image_to_video.transformer, EnhanceAVideoConfig(
|
115 |
+
# weight=config.enhance_a_video_weight if config.enable_enhance_a_video else 0.0,
|
116 |
+
# num_frames_callback=lambda: (config.num_frames - 1),
|
117 |
+
# _attention_type=1
|
118 |
+
# ))
|
119 |
else:
|
120 |
# Initialize text-to-video pipeline
|
121 |
self.text_to_video = HunyuanVideoPipeline.from_pretrained(
|
|
|
130 |
self.text_to_video.transformer = self.text_to_video.transformer.to(torch.bfloat16)
|
131 |
self.text_to_video.vae = self.text_to_video.vae.half()
|
132 |
|
133 |
+
|
134 |
+
# apply_enhance_a_video(self.text_to_video.transformer, EnhanceAVideoConfig(
|
135 |
+
# # weight=config.enhance_a_video_weight if config.enable_enhance_a_video else 0.0,
|
136 |
+
# weight=config.enhance_a_video_weight,
|
137 |
+
# num_frames_callback=lambda: (config.num_frames - 1),
|
138 |
+
# _attention_type=1
|
139 |
+
# ))
|
140 |
+
|
141 |
# enable FasterCache
|
142 |
|
143 |
# those values are coming from here:
|
|
|
350 |
config.input_image_quality,
|
351 |
)
|
352 |
generation_kwargs["image"] = processed_image
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
353 |
frames = self.image_to_video(**generation_kwargs).frames
|
354 |
else:
|
|
|
|
|
|
|
|
|
|
|
|
|
355 |
frames = self.text_to_video(**generation_kwargs).frames
|
356 |
|
357 |
|