Muhammad Taqi Raza commited on
Commit
e66051f
·
1 Parent(s): 33667dd

adding some prints

Browse files
cogvideo_controlnet_pcd.py CHANGED
@@ -172,8 +172,9 @@ class CogVideoXControlnetPCD(ModelMixin, ConfigMixin, PeftAdapterMixin):
172
  ):
173
  print("hidden_states.shape =", hidden_states.shape)
174
  print("controlnet_states.shape =", controlnet_states.shape)
175
- print("image_rotary_emb.shape =", image_rotary_emb.shape)
176
-
 
177
  hidden_states = torch.cat([hidden_states, controlnet_states], dim=2)
178
 
179
  # controlnet_states = self.controlnext_encoder(controlnet_states, timestep=timestep)
 
172
  ):
173
  print("hidden_states.shape =", hidden_states.shape)
174
  print("controlnet_states.shape =", controlnet_states.shape)
175
+ for i, emb in enumerate(image_rotary_emb):
176
+ print(f"image_rotary_emb[{i}].shape =", emb.shape)
177
+
178
  hidden_states = torch.cat([hidden_states, controlnet_states], dim=2)
179
 
180
  # controlnet_states = self.controlnext_encoder(controlnet_states, timestep=timestep)
scripts/inference.sh CHANGED
@@ -28,7 +28,7 @@ CUDA_VISIBLE_DEVICES=0 python inference/cli_demo_camera_i2v_pcd.py \
28
  --controlnet_transformer_out_proj_dim_zero_init \
29
  --vae_channels 16 \
30
  --num_frames 49 \
31
- --controlnet_transformer_num_layers 8 \
32
  --infer_with_mask \
33
  --pool_style 'max' \
34
  --seed 1
 
28
  --controlnet_transformer_out_proj_dim_zero_init \
29
  --vae_channels 16 \
30
  --num_frames 49 \
31
+ --controlnet_transformer_num_layers 8 \ # 4
32
  --infer_with_mask \
33
  --pool_style 'max' \
34
  --seed 1