1inkusFace commited on
Commit
71cad07
·
verified ·
1 Parent(s): 679d53d

Update pipeline_stable_diffusion_3_ipa.py

Browse files
Files changed (1) hide show
  1. pipeline_stable_diffusion_3_ipa.py +8 -5
pipeline_stable_diffusion_3_ipa.py CHANGED
@@ -1196,14 +1196,17 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
1196
  clip_image_embeds_5 = clip_image_embeds_5 * scale_5
1197
  image_prompt_embeds_list.append(clip_image_embeds_5)
1198
 
1199
- clip_image_embeds_cat_list = torch.cat(image_prompt_embeds_list).mean(dim=0)
1200
- print('catted embeds list with mean and unsqueeze: ',clip_image_embeds_cat_list.shape)
1201
- seq_len, _ = clip_image_embeds_cat_list.shape
1202
- clip_image_embeds_cat_list_repeat = clip_image_embeds_cat_list.repeat(1, 1, 1)
 
 
1203
  print('catted embeds repeat: ',clip_image_embeds_cat_list_repeat.shape)
1204
  clip_image_embeds_view = clip_image_embeds_cat_list_repeat.view(1, seq_len, -1)
1205
  print('catted viewed: ',clip_image_embeds_view.shape)
1206
- zeros_tensor = torch.zeros_like(clip_image_embeds_view)
 
1207
  print('zeros: ',zeros_tensor.shape)
1208
  clip_image_embeds = torch.cat([zeros_tensor, clip_image_embeds_view], dim=0)
1209
  print('embeds shape: ', clip_image_embeds.shape)
 
1196
  clip_image_embeds_5 = clip_image_embeds_5 * scale_5
1197
  image_prompt_embeds_list.append(clip_image_embeds_5)
1198
 
1199
+ #clip_image_embeds_cat_list = torch.cat(image_prompt_embeds_list).mean(dim=0)
1200
+ clip_image_embeds_cat_list = torch.cat(image_prompt_embeds_list)
1201
+ print('catted embeds list with mean: ',clip_image_embeds_cat_list.shape)
1202
+ #seq_len, _ = clip_image_embeds_cat_list.shape
1203
+ #clip_image_embeds_cat_list_repeat = clip_image_embeds_cat_list.repeat(1, 1, 1)
1204
+ clip_image_embeds_view = clip_image_embeds_cat.unsqueeze(0) # Added unsqueeze here instead
1205
  print('catted embeds repeat: ',clip_image_embeds_cat_list_repeat.shape)
1206
  clip_image_embeds_view = clip_image_embeds_cat_list_repeat.view(1, seq_len, -1)
1207
  print('catted viewed: ',clip_image_embeds_view.shape)
1208
+ #zeros_tensor = torch.zeros_like(clip_image_embeds_view)
1209
+ zeros_tensor = torch.zeros_like(clip_image_embeds_view[:, :image_prompt_embeds_list[0].shape[1], :]) # Make zeros tensor match the sequence length of a single image embedding
1210
  print('zeros: ',zeros_tensor.shape)
1211
  clip_image_embeds = torch.cat([zeros_tensor, clip_image_embeds_view], dim=0)
1212
  print('embeds shape: ', clip_image_embeds.shape)