Muhammad Taqi Raza
commited on
Commit
·
e66051f
1
Parent(s):
33667dd
adding some prints
Browse files- cogvideo_controlnet_pcd.py +3 -2
- scripts/inference.sh +1 -1
cogvideo_controlnet_pcd.py
CHANGED
@@ -172,8 +172,9 @@ class CogVideoXControlnetPCD(ModelMixin, ConfigMixin, PeftAdapterMixin):
|
|
172 |
):
|
173 |
print("hidden_states.shape =", hidden_states.shape)
|
174 |
print("controlnet_states.shape =", controlnet_states.shape)
|
175 |
-
|
176 |
-
|
|
|
177 |
hidden_states = torch.cat([hidden_states, controlnet_states], dim=2)
|
178 |
|
179 |
# controlnet_states = self.controlnext_encoder(controlnet_states, timestep=timestep)
|
|
|
172 |
):
|
173 |
print("hidden_states.shape =", hidden_states.shape)
|
174 |
print("controlnet_states.shape =", controlnet_states.shape)
|
175 |
+
for i, emb in enumerate(image_rotary_emb):
|
176 |
+
print(f"image_rotary_emb[{i}].shape =", emb.shape)
|
177 |
+
|
178 |
hidden_states = torch.cat([hidden_states, controlnet_states], dim=2)
|
179 |
|
180 |
# controlnet_states = self.controlnext_encoder(controlnet_states, timestep=timestep)
|
scripts/inference.sh
CHANGED
@@ -28,7 +28,7 @@ CUDA_VISIBLE_DEVICES=0 python inference/cli_demo_camera_i2v_pcd.py \
|
|
28 |
--controlnet_transformer_out_proj_dim_zero_init \
|
29 |
--vae_channels 16 \
|
30 |
--num_frames 49 \
|
31 |
-
--controlnet_transformer_num_layers 8 \
|
32 |
--infer_with_mask \
|
33 |
--pool_style 'max' \
|
34 |
--seed 1
|
|
|
28 |
--controlnet_transformer_out_proj_dim_zero_init \
|
29 |
--vae_channels 16 \
|
30 |
--num_frames 49 \
|
31 |
+
--controlnet_transformer_num_layers 8 \ # 4
|
32 |
--infer_with_mask \
|
33 |
--pool_style 'max' \
|
34 |
--seed 1
|