Datasets:

ArXiv:
diffusers-benchmarking-bot commited on
Commit
d89c14c
·
verified ·
1 Parent(s): 3831fbd

Upload folder using huggingface_hub

Browse files
Files changed (45) hide show
  1. main/adaptive_mask_inpainting.py +1 -1
  2. main/composable_stable_diffusion.py +1 -1
  3. main/edict_pipeline.py +1 -1
  4. main/fresco_v2v.py +1 -1
  5. main/gluegen.py +1 -1
  6. main/instaflow_one_step.py +1 -1
  7. main/ip_adapter_face_id.py +1 -1
  8. main/kohya_hires_fix.py +1 -1
  9. main/latent_consistency_img2img.py +1 -1
  10. main/latent_consistency_interpolate.py +1 -1
  11. main/latent_consistency_txt2img.py +1 -1
  12. main/llm_grounded_diffusion.py +1 -1
  13. main/lpw_stable_diffusion.py +1 -1
  14. main/lpw_stable_diffusion_xl.py +1 -1
  15. main/pipeline_animatediff_controlnet.py +1 -1
  16. main/pipeline_animatediff_img2video.py +1 -1
  17. main/pipeline_animatediff_ipex.py +1 -1
  18. main/pipeline_demofusion_sdxl.py +1 -1
  19. main/pipeline_fabric.py +1 -1
  20. main/pipeline_flux_differential_img2img.py +3 -4
  21. main/pipeline_flux_rf_inversion.py +1 -3
  22. main/pipeline_flux_with_cfg.py +1 -3
  23. main/pipeline_hunyuandit_differential_img2img.py +1 -3
  24. main/pipeline_kolors_differential_img2img.py +1 -3
  25. main/pipeline_prompt2prompt.py +1 -1
  26. main/pipeline_sdxl_style_aligned.py +1 -1
  27. main/pipeline_stable_diffusion_3_differential_img2img.py +1 -1
  28. main/pipeline_stable_diffusion_boxdiff.py +1 -1
  29. main/pipeline_stable_diffusion_pag.py +1 -1
  30. main/pipeline_stable_diffusion_upscale_ldm3d.py +1 -1
  31. main/pipeline_stable_diffusion_xl_controlnet_adapter.py +1 -1
  32. main/pipeline_stable_diffusion_xl_controlnet_adapter_inpaint.py +1 -1
  33. main/pipeline_stable_diffusion_xl_differential_img2img.py +1 -1
  34. main/pipeline_stable_diffusion_xl_ipex.py +1 -1
  35. main/pipeline_zero1to3.py +1 -1
  36. main/rerender_a_video.py +1 -1
  37. main/stable_diffusion_controlnet_img2img.py +1 -1
  38. main/stable_diffusion_controlnet_inpaint.py +1 -1
  39. main/stable_diffusion_controlnet_inpaint_img2img.py +1 -1
  40. main/stable_diffusion_ipex.py +1 -1
  41. main/stable_diffusion_reference.py +1 -1
  42. main/stable_diffusion_repaint.py +1 -1
  43. main/stable_diffusion_tensorrt_img2img.py +1 -1
  44. main/stable_diffusion_tensorrt_inpaint.py +1 -1
  45. main/stable_diffusion_tensorrt_txt2img.py +1 -1
main/adaptive_mask_inpainting.py CHANGED
@@ -450,7 +450,7 @@ class AdaptiveMaskInpaintPipeline(
450
  safety_checker=safety_checker,
451
  feature_extractor=feature_extractor,
452
  )
453
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
454
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
455
  self.register_to_config(requires_safety_checker=requires_safety_checker)
456
 
 
450
  safety_checker=safety_checker,
451
  feature_extractor=feature_extractor,
452
  )
453
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
454
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
455
  self.register_to_config(requires_safety_checker=requires_safety_checker)
456
 
main/composable_stable_diffusion.py CHANGED
@@ -162,7 +162,7 @@ class ComposableStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin)
162
  safety_checker=safety_checker,
163
  feature_extractor=feature_extractor,
164
  )
165
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
166
  self.register_to_config(requires_safety_checker=requires_safety_checker)
167
 
168
  def _encode_prompt(self, prompt, device, num_images_per_prompt, do_classifier_free_guidance, negative_prompt):
 
162
  safety_checker=safety_checker,
163
  feature_extractor=feature_extractor,
164
  )
165
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
166
  self.register_to_config(requires_safety_checker=requires_safety_checker)
167
 
168
  def _encode_prompt(self, prompt, device, num_images_per_prompt, do_classifier_free_guidance, negative_prompt):
main/edict_pipeline.py CHANGED
@@ -35,7 +35,7 @@ class EDICTPipeline(DiffusionPipeline):
35
  scheduler=scheduler,
36
  )
37
 
38
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
39
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
40
 
41
  def _encode_prompt(
 
35
  scheduler=scheduler,
36
  )
37
 
38
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
39
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
40
 
41
  def _encode_prompt(
main/fresco_v2v.py CHANGED
@@ -1342,7 +1342,7 @@ class FrescoV2VPipeline(StableDiffusionControlNetImg2ImgPipeline):
1342
  feature_extractor=feature_extractor,
1343
  image_encoder=image_encoder,
1344
  )
1345
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
1346
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True)
1347
  self.control_image_processor = VaeImageProcessor(
1348
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
 
1342
  feature_extractor=feature_extractor,
1343
  image_encoder=image_encoder,
1344
  )
1345
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
1346
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True)
1347
  self.control_image_processor = VaeImageProcessor(
1348
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
main/gluegen.py CHANGED
@@ -221,7 +221,7 @@ class GlueGenStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin, St
221
  language_adapter=language_adapter,
222
  tensor_norm=tensor_norm,
223
  )
224
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
225
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
226
  self.register_to_config(requires_safety_checker=requires_safety_checker)
227
 
 
221
  language_adapter=language_adapter,
222
  tensor_norm=tensor_norm,
223
  )
224
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
225
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
226
  self.register_to_config(requires_safety_checker=requires_safety_checker)
227
 
main/instaflow_one_step.py CHANGED
@@ -182,7 +182,7 @@ class InstaFlowPipeline(
182
  safety_checker=safety_checker,
183
  feature_extractor=feature_extractor,
184
  )
185
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
186
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
187
  self.register_to_config(requires_safety_checker=requires_safety_checker)
188
 
 
182
  safety_checker=safety_checker,
183
  feature_extractor=feature_extractor,
184
  )
185
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
186
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
187
  self.register_to_config(requires_safety_checker=requires_safety_checker)
188
 
main/ip_adapter_face_id.py CHANGED
@@ -265,7 +265,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(
265
  feature_extractor=feature_extractor,
266
  image_encoder=image_encoder,
267
  )
268
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
269
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
270
  self.register_to_config(requires_safety_checker=requires_safety_checker)
271
 
 
265
  feature_extractor=feature_extractor,
266
  image_encoder=image_encoder,
267
  )
268
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
269
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
270
  self.register_to_config(requires_safety_checker=requires_safety_checker)
271
 
main/kohya_hires_fix.py CHANGED
@@ -463,6 +463,6 @@ class StableDiffusionHighResFixPipeline(StableDiffusionPipeline):
463
  feature_extractor=feature_extractor,
464
  image_encoder=image_encoder,
465
  )
466
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
467
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
468
  self.register_to_config(requires_safety_checker=requires_safety_checker)
 
463
  feature_extractor=feature_extractor,
464
  image_encoder=image_encoder,
465
  )
466
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
467
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
468
  self.register_to_config(requires_safety_checker=requires_safety_checker)
main/latent_consistency_img2img.py CHANGED
@@ -69,7 +69,7 @@ class LatentConsistencyModelImg2ImgPipeline(DiffusionPipeline):
69
  safety_checker=safety_checker,
70
  feature_extractor=feature_extractor,
71
  )
72
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
73
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
74
 
75
  def _encode_prompt(
 
69
  safety_checker=safety_checker,
70
  feature_extractor=feature_extractor,
71
  )
72
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
73
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
74
 
75
  def _encode_prompt(
main/latent_consistency_interpolate.py CHANGED
@@ -273,7 +273,7 @@ class LatentConsistencyModelWalkPipeline(
273
  safety_checker=safety_checker,
274
  feature_extractor=feature_extractor,
275
  )
276
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
277
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
278
  self.register_to_config(requires_safety_checker=requires_safety_checker)
279
 
 
273
  safety_checker=safety_checker,
274
  feature_extractor=feature_extractor,
275
  )
276
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
277
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
278
  self.register_to_config(requires_safety_checker=requires_safety_checker)
279
 
main/latent_consistency_txt2img.py CHANGED
@@ -67,7 +67,7 @@ class LatentConsistencyModelPipeline(DiffusionPipeline):
67
  safety_checker=safety_checker,
68
  feature_extractor=feature_extractor,
69
  )
70
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
71
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
72
 
73
  def _encode_prompt(
 
67
  safety_checker=safety_checker,
68
  feature_extractor=feature_extractor,
69
  )
70
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
71
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
72
 
73
  def _encode_prompt(
main/llm_grounded_diffusion.py CHANGED
@@ -410,7 +410,7 @@ class LLMGroundedDiffusionPipeline(
410
  feature_extractor=feature_extractor,
411
  image_encoder=image_encoder,
412
  )
413
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
414
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
415
  self.register_to_config(requires_safety_checker=requires_safety_checker)
416
 
 
410
  feature_extractor=feature_extractor,
411
  image_encoder=image_encoder,
412
  )
413
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
414
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
415
  self.register_to_config(requires_safety_checker=requires_safety_checker)
416
 
main/lpw_stable_diffusion.py CHANGED
@@ -568,7 +568,7 @@ class StableDiffusionLongPromptWeightingPipeline(
568
  safety_checker=safety_checker,
569
  feature_extractor=feature_extractor,
570
  )
571
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
572
 
573
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
574
  self.register_to_config(
 
568
  safety_checker=safety_checker,
569
  feature_extractor=feature_extractor,
570
  )
571
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
572
 
573
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
574
  self.register_to_config(
main/lpw_stable_diffusion_xl.py CHANGED
@@ -673,7 +673,7 @@ class SDXLLongPromptWeightingPipeline(
673
  image_encoder=image_encoder,
674
  )
675
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
676
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
677
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
678
  self.mask_processor = VaeImageProcessor(
679
  vae_scale_factor=self.vae_scale_factor, do_normalize=False, do_binarize=True, do_convert_grayscale=True
 
673
  image_encoder=image_encoder,
674
  )
675
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
676
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
677
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
678
  self.mask_processor = VaeImageProcessor(
679
  vae_scale_factor=self.vae_scale_factor, do_normalize=False, do_binarize=True, do_convert_grayscale=True
main/pipeline_animatediff_controlnet.py CHANGED
@@ -188,7 +188,7 @@ class AnimateDiffControlNetPipeline(
188
  feature_extractor=feature_extractor,
189
  image_encoder=image_encoder,
190
  )
191
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
192
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
193
  self.control_image_processor = VaeImageProcessor(
194
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
 
188
  feature_extractor=feature_extractor,
189
  image_encoder=image_encoder,
190
  )
191
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
192
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
193
  self.control_image_processor = VaeImageProcessor(
194
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
main/pipeline_animatediff_img2video.py CHANGED
@@ -308,7 +308,7 @@ class AnimateDiffImgToVideoPipeline(
308
  feature_extractor=feature_extractor,
309
  image_encoder=image_encoder,
310
  )
311
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
312
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
313
 
314
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline.encode_prompt with num_images_per_prompt -> num_videos_per_prompt
 
308
  feature_extractor=feature_extractor,
309
  image_encoder=image_encoder,
310
  )
311
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
312
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
313
 
314
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline.encode_prompt with num_images_per_prompt -> num_videos_per_prompt
main/pipeline_animatediff_ipex.py CHANGED
@@ -162,7 +162,7 @@ class AnimateDiffPipelineIpex(
162
  feature_extractor=feature_extractor,
163
  image_encoder=image_encoder,
164
  )
165
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
166
  self.video_processor = VideoProcessor(do_resize=False, vae_scale_factor=self.vae_scale_factor)
167
 
168
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline.encode_prompt with num_images_per_prompt -> num_videos_per_prompt
 
162
  feature_extractor=feature_extractor,
163
  image_encoder=image_encoder,
164
  )
165
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
166
  self.video_processor = VideoProcessor(do_resize=False, vae_scale_factor=self.vae_scale_factor)
167
 
168
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline.encode_prompt with num_images_per_prompt -> num_videos_per_prompt
main/pipeline_demofusion_sdxl.py CHANGED
@@ -166,7 +166,7 @@ class DemoFusionSDXLPipeline(
166
  scheduler=scheduler,
167
  )
168
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
169
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
170
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
171
  self.default_sample_size = self.unet.config.sample_size
172
 
 
166
  scheduler=scheduler,
167
  )
168
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
169
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
170
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
171
  self.default_sample_size = self.unet.config.sample_size
172
 
main/pipeline_fabric.py CHANGED
@@ -179,7 +179,7 @@ class FabricPipeline(DiffusionPipeline):
179
  tokenizer=tokenizer,
180
  scheduler=scheduler,
181
  )
182
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
183
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
184
 
185
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._encode_prompt
 
179
  tokenizer=tokenizer,
180
  scheduler=scheduler,
181
  )
182
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
183
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
184
 
185
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._encode_prompt
main/pipeline_flux_differential_img2img.py CHANGED
@@ -221,13 +221,12 @@ class FluxDifferentialImg2ImgPipeline(DiffusionPipeline, FluxLoraLoaderMixin):
221
  transformer=transformer,
222
  scheduler=scheduler,
223
  )
224
- self.vae_scale_factor = (
225
- 2 ** (len(self.vae.config.block_out_channels)) if hasattr(self, "vae") and self.vae is not None else 16
226
- )
227
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
 
228
  self.mask_processor = VaeImageProcessor(
229
  vae_scale_factor=self.vae_scale_factor,
230
- vae_latent_channels=self.vae.config.latent_channels,
231
  do_normalize=False,
232
  do_binarize=False,
233
  do_convert_grayscale=True,
 
221
  transformer=transformer,
222
  scheduler=scheduler,
223
  )
224
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels)) if getattr(self, "vae", None) else 16
 
 
225
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
226
+ latent_channels = self.vae.config.latent_channels if getattr(self, "vae", None) else 16
227
  self.mask_processor = VaeImageProcessor(
228
  vae_scale_factor=self.vae_scale_factor,
229
+ vae_latent_channels=latent_channels,
230
  do_normalize=False,
231
  do_binarize=False,
232
  do_convert_grayscale=True,
main/pipeline_flux_rf_inversion.py CHANGED
@@ -219,9 +219,7 @@ class RFInversionFluxPipeline(
219
  transformer=transformer,
220
  scheduler=scheduler,
221
  )
222
- self.vae_scale_factor = (
223
- 2 ** (len(self.vae.config.block_out_channels) - 1) if hasattr(self, "vae") and self.vae is not None else 8
224
- )
225
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
226
  self.tokenizer_max_length = (
227
  self.tokenizer.model_max_length if hasattr(self, "tokenizer") and self.tokenizer is not None else 77
 
219
  transformer=transformer,
220
  scheduler=scheduler,
221
  )
222
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
 
 
223
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
224
  self.tokenizer_max_length = (
225
  self.tokenizer.model_max_length if hasattr(self, "tokenizer") and self.tokenizer is not None else 77
main/pipeline_flux_with_cfg.py CHANGED
@@ -189,9 +189,7 @@ class FluxCFGPipeline(DiffusionPipeline, FluxLoraLoaderMixin, FromSingleFileMixi
189
  transformer=transformer,
190
  scheduler=scheduler,
191
  )
192
- self.vae_scale_factor = (
193
- 2 ** (len(self.vae.config.block_out_channels)) if hasattr(self, "vae") and self.vae is not None else 16
194
- )
195
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
196
  self.tokenizer_max_length = (
197
  self.tokenizer.model_max_length if hasattr(self, "tokenizer") and self.tokenizer is not None else 77
 
189
  transformer=transformer,
190
  scheduler=scheduler,
191
  )
192
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels)) if getattr(self, "vae", None) else 16
 
 
193
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
194
  self.tokenizer_max_length = (
195
  self.tokenizer.model_max_length if hasattr(self, "tokenizer") and self.tokenizer is not None else 77
main/pipeline_hunyuandit_differential_img2img.py CHANGED
@@ -327,9 +327,7 @@ class HunyuanDiTDifferentialImg2ImgPipeline(DiffusionPipeline):
327
  " checker. If you do not want to use the safety checker, you can pass `'safety_checker=None'` instead."
328
  )
329
 
330
- self.vae_scale_factor = (
331
- 2 ** (len(self.vae.config.block_out_channels) - 1) if hasattr(self, "vae") and self.vae is not None else 8
332
- )
333
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
334
  self.mask_processor = VaeImageProcessor(
335
  vae_scale_factor=self.vae_scale_factor,
 
327
  " checker. If you do not want to use the safety checker, you can pass `'safety_checker=None'` instead."
328
  )
329
 
330
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
 
 
331
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
332
  self.mask_processor = VaeImageProcessor(
333
  vae_scale_factor=self.vae_scale_factor,
main/pipeline_kolors_differential_img2img.py CHANGED
@@ -209,9 +209,7 @@ class KolorsDifferentialImg2ImgPipeline(
209
  feature_extractor=feature_extractor,
210
  )
211
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
212
- self.vae_scale_factor = (
213
- 2 ** (len(self.vae.config.block_out_channels) - 1) if hasattr(self, "vae") and self.vae is not None else 8
214
- )
215
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
216
 
217
  self.mask_processor = VaeImageProcessor(
 
209
  feature_extractor=feature_extractor,
210
  )
211
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
212
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
 
 
213
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
214
 
215
  self.mask_processor = VaeImageProcessor(
main/pipeline_prompt2prompt.py CHANGED
@@ -205,7 +205,7 @@ class Prompt2PromptPipeline(
205
  feature_extractor=feature_extractor,
206
  image_encoder=image_encoder,
207
  )
208
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
209
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
210
  self.register_to_config(requires_safety_checker=requires_safety_checker)
211
 
 
205
  feature_extractor=feature_extractor,
206
  image_encoder=image_encoder,
207
  )
208
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
209
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
210
  self.register_to_config(requires_safety_checker=requires_safety_checker)
211
 
main/pipeline_sdxl_style_aligned.py CHANGED
@@ -488,7 +488,7 @@ class StyleAlignedSDXLPipeline(
488
  feature_extractor=feature_extractor,
489
  )
490
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
491
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
492
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
493
  self.mask_processor = VaeImageProcessor(
494
  vae_scale_factor=self.vae_scale_factor, do_normalize=False, do_binarize=True, do_convert_grayscale=True
 
488
  feature_extractor=feature_extractor,
489
  )
490
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
491
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
492
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
493
  self.mask_processor = VaeImageProcessor(
494
  vae_scale_factor=self.vae_scale_factor, do_normalize=False, do_binarize=True, do_convert_grayscale=True
main/pipeline_stable_diffusion_3_differential_img2img.py CHANGED
@@ -207,7 +207,7 @@ class StableDiffusion3DifferentialImg2ImgPipeline(DiffusionPipeline):
207
  transformer=transformer,
208
  scheduler=scheduler,
209
  )
210
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
211
  self.image_processor = VaeImageProcessor(
212
  vae_scale_factor=self.vae_scale_factor, vae_latent_channels=self.vae.config.latent_channels
213
  )
 
207
  transformer=transformer,
208
  scheduler=scheduler,
209
  )
210
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
211
  self.image_processor = VaeImageProcessor(
212
  vae_scale_factor=self.vae_scale_factor, vae_latent_channels=self.vae.config.latent_channels
213
  )
main/pipeline_stable_diffusion_boxdiff.py CHANGED
@@ -491,7 +491,7 @@ class StableDiffusionBoxDiffPipeline(
491
  feature_extractor=feature_extractor,
492
  image_encoder=image_encoder,
493
  )
494
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
495
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
496
  self.register_to_config(requires_safety_checker=requires_safety_checker)
497
 
 
491
  feature_extractor=feature_extractor,
492
  image_encoder=image_encoder,
493
  )
494
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
495
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
496
  self.register_to_config(requires_safety_checker=requires_safety_checker)
497
 
main/pipeline_stable_diffusion_pag.py CHANGED
@@ -458,7 +458,7 @@ class StableDiffusionPAGPipeline(
458
  feature_extractor=feature_extractor,
459
  image_encoder=image_encoder,
460
  )
461
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
462
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
463
  self.register_to_config(requires_safety_checker=requires_safety_checker)
464
 
 
458
  feature_extractor=feature_extractor,
459
  image_encoder=image_encoder,
460
  )
461
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
462
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
463
  self.register_to_config(requires_safety_checker=requires_safety_checker)
464
 
main/pipeline_stable_diffusion_upscale_ldm3d.py CHANGED
@@ -151,7 +151,7 @@ class StableDiffusionUpscaleLDM3DPipeline(
151
  watermarker=watermarker,
152
  feature_extractor=feature_extractor,
153
  )
154
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
155
  self.image_processor = VaeImageProcessorLDM3D(vae_scale_factor=self.vae_scale_factor, resample="bilinear")
156
  # self.register_to_config(requires_safety_checker=requires_safety_checker)
157
  self.register_to_config(max_noise_level=max_noise_level)
 
151
  watermarker=watermarker,
152
  feature_extractor=feature_extractor,
153
  )
154
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
155
  self.image_processor = VaeImageProcessorLDM3D(vae_scale_factor=self.vae_scale_factor, resample="bilinear")
156
  # self.register_to_config(requires_safety_checker=requires_safety_checker)
157
  self.register_to_config(max_noise_level=max_noise_level)
main/pipeline_stable_diffusion_xl_controlnet_adapter.py CHANGED
@@ -226,7 +226,7 @@ class StableDiffusionXLControlNetAdapterPipeline(
226
  scheduler=scheduler,
227
  )
228
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
229
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
230
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
231
  self.control_image_processor = VaeImageProcessor(
232
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
 
226
  scheduler=scheduler,
227
  )
228
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
229
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
230
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
231
  self.control_image_processor = VaeImageProcessor(
232
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
main/pipeline_stable_diffusion_xl_controlnet_adapter_inpaint.py CHANGED
@@ -374,7 +374,7 @@ class StableDiffusionXLControlNetAdapterInpaintPipeline(
374
  )
375
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
376
  self.register_to_config(requires_aesthetics_score=requires_aesthetics_score)
377
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
378
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
379
  self.control_image_processor = VaeImageProcessor(
380
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
 
374
  )
375
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
376
  self.register_to_config(requires_aesthetics_score=requires_aesthetics_score)
377
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
378
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
379
  self.control_image_processor = VaeImageProcessor(
380
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
main/pipeline_stable_diffusion_xl_differential_img2img.py CHANGED
@@ -258,7 +258,7 @@ class StableDiffusionXLDifferentialImg2ImgPipeline(
258
  )
259
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
260
  self.register_to_config(requires_aesthetics_score=requires_aesthetics_score)
261
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
262
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
263
 
264
  add_watermarker = add_watermarker if add_watermarker is not None else is_invisible_watermark_available()
 
258
  )
259
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
260
  self.register_to_config(requires_aesthetics_score=requires_aesthetics_score)
261
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
262
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
263
 
264
  add_watermarker = add_watermarker if add_watermarker is not None else is_invisible_watermark_available()
main/pipeline_stable_diffusion_xl_ipex.py CHANGED
@@ -253,7 +253,7 @@ class StableDiffusionXLPipelineIpex(
253
  feature_extractor=feature_extractor,
254
  )
255
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
256
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
257
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
258
 
259
  self.default_sample_size = self.unet.config.sample_size
 
253
  feature_extractor=feature_extractor,
254
  )
255
  self.register_to_config(force_zeros_for_empty_prompt=force_zeros_for_empty_prompt)
256
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
257
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
258
 
259
  self.default_sample_size = self.unet.config.sample_size
main/pipeline_zero1to3.py CHANGED
@@ -181,7 +181,7 @@ class Zero1to3StableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin):
181
  feature_extractor=feature_extractor,
182
  cc_projection=cc_projection,
183
  )
184
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
185
  self.register_to_config(requires_safety_checker=requires_safety_checker)
186
  # self.model_mode = None
187
 
 
181
  feature_extractor=feature_extractor,
182
  cc_projection=cc_projection,
183
  )
184
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
185
  self.register_to_config(requires_safety_checker=requires_safety_checker)
186
  # self.model_mode = None
187
 
main/rerender_a_video.py CHANGED
@@ -352,7 +352,7 @@ class RerenderAVideoPipeline(StableDiffusionControlNetImg2ImgPipeline):
352
  safety_checker=safety_checker,
353
  feature_extractor=feature_extractor,
354
  )
355
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
356
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True)
357
  self.control_image_processor = VaeImageProcessor(
358
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
 
352
  safety_checker=safety_checker,
353
  feature_extractor=feature_extractor,
354
  )
355
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
356
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True)
357
  self.control_image_processor = VaeImageProcessor(
358
  vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False
main/stable_diffusion_controlnet_img2img.py CHANGED
@@ -179,7 +179,7 @@ class StableDiffusionControlNetImg2ImgPipeline(DiffusionPipeline, StableDiffusio
179
  safety_checker=safety_checker,
180
  feature_extractor=feature_extractor,
181
  )
182
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
183
  self.register_to_config(requires_safety_checker=requires_safety_checker)
184
 
185
  def _encode_prompt(
 
179
  safety_checker=safety_checker,
180
  feature_extractor=feature_extractor,
181
  )
182
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
183
  self.register_to_config(requires_safety_checker=requires_safety_checker)
184
 
185
  def _encode_prompt(
main/stable_diffusion_controlnet_inpaint.py CHANGED
@@ -278,7 +278,7 @@ class StableDiffusionControlNetInpaintPipeline(DiffusionPipeline, StableDiffusio
278
  feature_extractor=feature_extractor,
279
  )
280
 
281
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
282
  self.register_to_config(requires_safety_checker=requires_safety_checker)
283
 
284
  def _encode_prompt(
 
278
  feature_extractor=feature_extractor,
279
  )
280
 
281
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
282
  self.register_to_config(requires_safety_checker=requires_safety_checker)
283
 
284
  def _encode_prompt(
main/stable_diffusion_controlnet_inpaint_img2img.py CHANGED
@@ -263,7 +263,7 @@ class StableDiffusionControlNetInpaintImg2ImgPipeline(DiffusionPipeline, StableD
263
  safety_checker=safety_checker,
264
  feature_extractor=feature_extractor,
265
  )
266
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
267
  self.register_to_config(requires_safety_checker=requires_safety_checker)
268
 
269
  def _encode_prompt(
 
263
  safety_checker=safety_checker,
264
  feature_extractor=feature_extractor,
265
  )
266
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
267
  self.register_to_config(requires_safety_checker=requires_safety_checker)
268
 
269
  def _encode_prompt(
main/stable_diffusion_ipex.py CHANGED
@@ -178,7 +178,7 @@ class StableDiffusionIPEXPipeline(
178
  safety_checker=safety_checker,
179
  feature_extractor=feature_extractor,
180
  )
181
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
182
  self.register_to_config(requires_safety_checker=requires_safety_checker)
183
 
184
  def get_input_example(self, prompt, height=None, width=None, guidance_scale=7.5, num_images_per_prompt=1):
 
178
  safety_checker=safety_checker,
179
  feature_extractor=feature_extractor,
180
  )
181
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
182
  self.register_to_config(requires_safety_checker=requires_safety_checker)
183
 
184
  def get_input_example(self, prompt, height=None, width=None, guidance_scale=7.5, num_images_per_prompt=1):
main/stable_diffusion_reference.py CHANGED
@@ -219,7 +219,7 @@ class StableDiffusionReferencePipeline(
219
  safety_checker=safety_checker,
220
  feature_extractor=feature_extractor,
221
  )
222
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
223
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
224
  self.register_to_config(requires_safety_checker=requires_safety_checker)
225
 
 
219
  safety_checker=safety_checker,
220
  feature_extractor=feature_extractor,
221
  )
222
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
223
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
224
  self.register_to_config(requires_safety_checker=requires_safety_checker)
225
 
main/stable_diffusion_repaint.py CHANGED
@@ -274,7 +274,7 @@ class StableDiffusionRepaintPipeline(
274
  safety_checker=safety_checker,
275
  feature_extractor=feature_extractor,
276
  )
277
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
278
  self.register_to_config(requires_safety_checker=requires_safety_checker)
279
 
280
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._encode_prompt
 
274
  safety_checker=safety_checker,
275
  feature_extractor=feature_extractor,
276
  )
277
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
278
  self.register_to_config(requires_safety_checker=requires_safety_checker)
279
 
280
  # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._encode_prompt
main/stable_diffusion_tensorrt_img2img.py CHANGED
@@ -806,7 +806,7 @@ class TensorRTStableDiffusionImg2ImgPipeline(DiffusionPipeline):
806
  self.engine = {} # loaded in build_engines()
807
 
808
  self.vae.forward = self.vae.decode
809
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
810
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
811
  self.register_to_config(requires_safety_checker=requires_safety_checker)
812
 
 
806
  self.engine = {} # loaded in build_engines()
807
 
808
  self.vae.forward = self.vae.decode
809
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
810
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
811
  self.register_to_config(requires_safety_checker=requires_safety_checker)
812
 
main/stable_diffusion_tensorrt_inpaint.py CHANGED
@@ -810,7 +810,7 @@ class TensorRTStableDiffusionInpaintPipeline(DiffusionPipeline):
810
  self.engine = {} # loaded in build_engines()
811
 
812
  self.vae.forward = self.vae.decode
813
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
814
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
815
  self.register_to_config(requires_safety_checker=requires_safety_checker)
816
 
 
810
  self.engine = {} # loaded in build_engines()
811
 
812
  self.vae.forward = self.vae.decode
813
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
814
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
815
  self.register_to_config(requires_safety_checker=requires_safety_checker)
816
 
main/stable_diffusion_tensorrt_txt2img.py CHANGED
@@ -722,7 +722,7 @@ class TensorRTStableDiffusionPipeline(DiffusionPipeline):
722
  self.engine = {} # loaded in build_engines()
723
 
724
  self.vae.forward = self.vae.decode
725
- self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1)
726
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
727
  self.register_to_config(requires_safety_checker=requires_safety_checker)
728
 
 
722
  self.engine = {} # loaded in build_engines()
723
 
724
  self.vae.forward = self.vae.decode
725
+ self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) if getattr(self, "vae", None) else 8
726
  self.image_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor)
727
  self.register_to_config(requires_safety_checker=requires_safety_checker)
728