Update pipeline_stable_diffusion_3_ipa.py
Browse files
pipeline_stable_diffusion_3_ipa.py
CHANGED
@@ -1186,7 +1186,12 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
|
|
1186 |
#clip_image_embeds = torch.cat([torch.zeros_like(torch.stack(image_prompt_embeds_list)), torch.stack(image_prompt_embeds_list)], dim=0).mean(dim=0)
|
1187 |
# FAILS clip_image_embeds = torch.cat(torch.stack(image_prompt_embeds_list), dim=0).mean(dim=0)
|
1188 |
# FAILS TIMESTEPS clip_image_embeds = torch.cat(image_prompt_embeds_list, dim=0).mean(dim=0)
|
1189 |
-
|
|
|
|
|
|
|
|
|
|
|
1190 |
#clip_image_embeds = torch.cat([torch.stack(image_prompt_embeds_list)], dim=0).mean(dim=0)
|
1191 |
|
1192 |
# 4. Prepare timesteps
|
|
|
1186 |
#clip_image_embeds = torch.cat([torch.zeros_like(torch.stack(image_prompt_embeds_list)), torch.stack(image_prompt_embeds_list)], dim=0).mean(dim=0)
|
1187 |
# FAILS clip_image_embeds = torch.cat(torch.stack(image_prompt_embeds_list), dim=0).mean(dim=0)
|
1188 |
# FAILS TIMESTEPS clip_image_embeds = torch.cat(image_prompt_embeds_list, dim=0).mean(dim=0)
|
1189 |
+
|
1190 |
+
clip_image_embeds = torch.mean(torch.stack(image_prompt_embeds_list), dim=0).unsqueeze(0)
|
1191 |
+
bs_embed, seq_len, _ = image_prompt_embeds.shape
|
1192 |
+
image_prompt_embeds = image_prompt_embeds.repeat(1, 1, 1)
|
1193 |
+
image_prompt_embeds = image_prompt_embeds.view(bs_embed * 1, seq_len, -1)
|
1194 |
+
|
1195 |
#clip_image_embeds = torch.cat([torch.stack(image_prompt_embeds_list)], dim=0).mean(dim=0)
|
1196 |
|
1197 |
# 4. Prepare timesteps
|