Update pipeline_stable_diffusion_3_ipa.py
Browse files
pipeline_stable_diffusion_3_ipa.py
CHANGED
@@ -1196,14 +1196,17 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
|
|
1196 |
clip_image_embeds_5 = clip_image_embeds_5 * scale_5
|
1197 |
image_prompt_embeds_list.append(clip_image_embeds_5)
|
1198 |
|
1199 |
-
clip_image_embeds_cat_list = torch.cat(image_prompt_embeds_list).mean(dim=0)
|
1200 |
-
|
1201 |
-
|
1202 |
-
|
|
|
|
|
1203 |
print('catted embeds repeat: ',clip_image_embeds_cat_list_repeat.shape)
|
1204 |
clip_image_embeds_view = clip_image_embeds_cat_list_repeat.view(1, seq_len, -1)
|
1205 |
print('catted viewed: ',clip_image_embeds_view.shape)
|
1206 |
-
zeros_tensor = torch.zeros_like(clip_image_embeds_view)
|
|
|
1207 |
print('zeros: ',zeros_tensor.shape)
|
1208 |
clip_image_embeds = torch.cat([zeros_tensor, clip_image_embeds_view], dim=0)
|
1209 |
print('embeds shape: ', clip_image_embeds.shape)
|
|
|
1196 |
clip_image_embeds_5 = clip_image_embeds_5 * scale_5
|
1197 |
image_prompt_embeds_list.append(clip_image_embeds_5)
|
1198 |
|
1199 |
+
#clip_image_embeds_cat_list = torch.cat(image_prompt_embeds_list).mean(dim=0)
|
1200 |
+
clip_image_embeds_cat_list = torch.cat(image_prompt_embeds_list)
|
1201 |
+
print('catted embeds list with mean: ',clip_image_embeds_cat_list.shape)
|
1202 |
+
#seq_len, _ = clip_image_embeds_cat_list.shape
|
1203 |
+
#clip_image_embeds_cat_list_repeat = clip_image_embeds_cat_list.repeat(1, 1, 1)
|
1204 |
+
clip_image_embeds_view = clip_image_embeds_cat.unsqueeze(0) # Added unsqueeze here instead
|
1205 |
print('catted embeds repeat: ',clip_image_embeds_cat_list_repeat.shape)
|
1206 |
clip_image_embeds_view = clip_image_embeds_cat_list_repeat.view(1, seq_len, -1)
|
1207 |
print('catted viewed: ',clip_image_embeds_view.shape)
|
1208 |
+
#zeros_tensor = torch.zeros_like(clip_image_embeds_view)
|
1209 |
+
zeros_tensor = torch.zeros_like(clip_image_embeds_view[:, :image_prompt_embeds_list[0].shape[1], :]) # Make zeros tensor match the sequence length of a single image embedding
|
1210 |
print('zeros: ',zeros_tensor.shape)
|
1211 |
clip_image_embeds = torch.cat([zeros_tensor, clip_image_embeds_view], dim=0)
|
1212 |
print('embeds shape: ', clip_image_embeds.shape)
|