Muhammad Taqi Raza commited on
Commit
23f29c8
·
1 Parent(s): 38298be

modifying requirements.txt

Browse files
inference/cli_demo_camera_i2v_pcd.py CHANGED
@@ -14,7 +14,7 @@ from diffusers import (
14
  AutoencoderKLCogVideoX
15
  )
16
  from diffusers.utils import export_to_video, load_video
17
-
18
  from controlnet_pipeline import ControlnetCogVideoXImageToVideoPCDPipeline
19
  from cogvideo_transformer import CustomCogVideoXTransformer3DModel
20
  from cogvideo_controlnet_pcd import CogVideoXControlnetPCD
@@ -369,8 +369,8 @@ def generate_video(
369
  # ++++++++++++++++++++++++++++++++++++++
370
  latents = video_generate_all # This is a latent
371
  # Fix the nested list structure
372
- if isinstance(latents, list) and len(latents) == 1 and isinstance(latents[0], list):
373
- latents = latents[0] # ✅ Unwrap the inner list of PIL images
374
 
375
  transform = T.ToTensor() # Converts PIL image to torch.FloatTensor in [0,1]
376
 
@@ -403,7 +403,13 @@ def generate_video(
403
  if rife_status:
404
  latents = rife_inference_with_latents(frame_interpolation_model, latents)
405
 
 
 
 
 
 
406
  video_generate_all = latents
 
407
  # ++++++++++++++++++++++++++++++++++++++
408
 
409
  video_generate = video_generate_all[0]
 
14
  AutoencoderKLCogVideoX
15
  )
16
  from diffusers.utils import export_to_video, load_video
17
+ from torchvision.transforms import ToPILImage
18
  from controlnet_pipeline import ControlnetCogVideoXImageToVideoPCDPipeline
19
  from cogvideo_transformer import CustomCogVideoXTransformer3DModel
20
  from cogvideo_controlnet_pcd import CogVideoXControlnetPCD
 
369
  # ++++++++++++++++++++++++++++++++++++++
370
  latents = video_generate_all # This is a latent
371
  # Fix the nested list structure
372
+ # if isinstance(latents, list) and len(latents) == 1 and isinstance(latents[0], list):
373
+ # latents = latents[0] # ✅ Unwrap the inner list of PIL images
374
 
375
  transform = T.ToTensor() # Converts PIL image to torch.FloatTensor in [0,1]
376
 
 
403
  if rife_status:
404
  latents = rife_inference_with_latents(frame_interpolation_model, latents)
405
 
406
+
407
+ # Convert latents back to PIL images after processing
408
+ latents = latents.clamp(0, 1) # Clamp values to [0,1]
409
+ latents = [T.ToPILImage()(frame.cpu()) for frame in latents]
410
+
411
  video_generate_all = latents
412
+
413
  # ++++++++++++++++++++++++++++++++++++++
414
 
415
  video_generate = video_generate_all[0]