Muhammad Taqi Raza
commited on
Commit
·
23f29c8
1
Parent(s):
38298be
modifying requirements.txt
Browse files
inference/cli_demo_camera_i2v_pcd.py
CHANGED
@@ -14,7 +14,7 @@ from diffusers import (
|
|
14 |
AutoencoderKLCogVideoX
|
15 |
)
|
16 |
from diffusers.utils import export_to_video, load_video
|
17 |
-
|
18 |
from controlnet_pipeline import ControlnetCogVideoXImageToVideoPCDPipeline
|
19 |
from cogvideo_transformer import CustomCogVideoXTransformer3DModel
|
20 |
from cogvideo_controlnet_pcd import CogVideoXControlnetPCD
|
@@ -369,8 +369,8 @@ def generate_video(
|
|
369 |
# ++++++++++++++++++++++++++++++++++++++
|
370 |
latents = video_generate_all # This is a latent
|
371 |
# Fix the nested list structure
|
372 |
-
if isinstance(latents, list) and len(latents) == 1 and isinstance(latents[0], list):
|
373 |
-
|
374 |
|
375 |
transform = T.ToTensor() # Converts PIL image to torch.FloatTensor in [0,1]
|
376 |
|
@@ -403,7 +403,13 @@ def generate_video(
|
|
403 |
if rife_status:
|
404 |
latents = rife_inference_with_latents(frame_interpolation_model, latents)
|
405 |
|
|
|
|
|
|
|
|
|
|
|
406 |
video_generate_all = latents
|
|
|
407 |
# ++++++++++++++++++++++++++++++++++++++
|
408 |
|
409 |
video_generate = video_generate_all[0]
|
|
|
14 |
AutoencoderKLCogVideoX
|
15 |
)
|
16 |
from diffusers.utils import export_to_video, load_video
|
17 |
+
from torchvision.transforms import ToPILImage
|
18 |
from controlnet_pipeline import ControlnetCogVideoXImageToVideoPCDPipeline
|
19 |
from cogvideo_transformer import CustomCogVideoXTransformer3DModel
|
20 |
from cogvideo_controlnet_pcd import CogVideoXControlnetPCD
|
|
|
369 |
# ++++++++++++++++++++++++++++++++++++++
|
370 |
latents = video_generate_all # This is a latent
|
371 |
# Fix the nested list structure
|
372 |
+
# if isinstance(latents, list) and len(latents) == 1 and isinstance(latents[0], list):
|
373 |
+
# latents = latents[0] # ✅ Unwrap the inner list of PIL images
|
374 |
|
375 |
transform = T.ToTensor() # Converts PIL image to torch.FloatTensor in [0,1]
|
376 |
|
|
|
403 |
if rife_status:
|
404 |
latents = rife_inference_with_latents(frame_interpolation_model, latents)
|
405 |
|
406 |
+
|
407 |
+
# Convert latents back to PIL images after processing
|
408 |
+
latents = latents.clamp(0, 1) # Clamp values to [0,1]
|
409 |
+
latents = [T.ToPILImage()(frame.cpu()) for frame in latents]
|
410 |
+
|
411 |
video_generate_all = latents
|
412 |
+
|
413 |
# ++++++++++++++++++++++++++++++++++++++
|
414 |
|
415 |
video_generate = video_generate_all[0]
|