Skip to content

Commit 4d5659d

Browse files
committed
Bugfix for dreambooth flux2 img2img2
1 parent d503783 commit 4d5659d

File tree

49 files changed

+35
-572
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

49 files changed

+35
-572
lines changed

src/diffusers/loaders/lora_pipeline.py

Lines changed: 18 additions & 31 deletions
Original file line numberDiff line numberDiff line change
@@ -1208,7 +1208,7 @@ def load_lora_into_text_encoder(
12081208
@classmethod
12091209
# Copied from diffusers.loaders.lora_pipeline.StableDiffusionXLLoraLoaderMixin.save_lora_weights with unet->transformer
12101210
def save_lora_weights(
1211-
cls,
1211+
cls,
12121212
save_directory: Union[str, os.PathLike],
12131213
transformer_lora_layers: Dict[str, Union[torch.nn.Module, torch.Tensor]] = None,
12141214
text_encoder_lora_layers: Dict[str, Union[torch.nn.Module, torch.Tensor]] = None,
@@ -1256,7 +1256,7 @@ def save_lora_weights(
12561256

12571257
# Copied from diffusers.loaders.lora_pipeline.StableDiffusionXLLoraLoaderMixin.fuse_lora with unet->transformer
12581258
def fuse_lora(
1259-
self,
1259+
self,
12601260
components: List[str] = ["transformer", "text_encoder", "text_encoder_2"],
12611261
lora_scale: float = 1.0,
12621262
safe_fusing: bool = False,
@@ -1276,7 +1276,7 @@ def fuse_lora(
12761276

12771277
# Copied from diffusers.loaders.lora_pipeline.StableDiffusionXLLoraLoaderMixin.unfuse_lora with unet->transformer
12781278
def unfuse_lora(self, components: List[str] = ["transformer", "text_encoder", "text_encoder_2"], **kwargs):
1279-
r"""
1279+
r"""
12801280
See [`~loaders.StableDiffusionLoraLoaderMixin.unfuse_lora`] for more details.
12811281
"""
12821282
super().unfuse_lora(components=components, **kwargs)
@@ -1390,7 +1390,7 @@ def load_lora_weights(
13901390
@classmethod
13911391
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->AuraFlowTransformer2DModel
13921392
def load_lora_into_transformer(
1393-
cls,
1393+
cls,
13941394
state_dict,
13951395
transformer,
13961396
adapter_name=None,
@@ -1893,19 +1893,6 @@ def load_lora_into_text_encoder(
18931893
@classmethod
18941894
# Copied from diffusers.loaders.lora_pipeline.StableDiffusionLoraLoaderMixin.save_lora_weights with unet->transformer
18951895
def save_lora_weights(
1896-
cls,
1897-
save_directory: Union[str, os.PathLike],
1898-
transformer_lora_layers: Dict[str, Union[torch.nn.Module, torch.Tensor]] = None,
1899-
text_encoder_lora_layers: Dict[str, torch.nn.Module] = None,
1900-
is_main_process: bool = True,
1901-
weight_name: str = None,
1902-
save_function: Callable = None,
1903-
safe_serialization: bool = True,
1904-
transformer_lora_adapter_metadata=None,
1905-
text_encoder_lora_adapter_metadata=None,
1906-
):
1907-
r"""
1908-
Save the LoRA parameters corresponding to the UNet and text encoder.
19091896

19101897
Arguments:
19111898
save_directory (`str` or `os.PathLike`):
@@ -2291,7 +2278,7 @@ class AmusedLoraLoaderMixin(StableDiffusionLoraLoaderMixin):
22912278
@classmethod
22922279
# Copied from diffusers.loaders.lora_pipeline.FluxLoraLoaderMixin.load_lora_into_transformer with FluxTransformer2DModel->UVit2DModel
22932280
def load_lora_into_transformer(
2294-
cls,
2281+
cls,
22952282
state_dict,
22962283
network_alphas,
22972284
transformer,
@@ -2542,7 +2529,7 @@ def load_lora_weights(
25422529
@classmethod
25432530
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->CogVideoXTransformer3DModel
25442531
def load_lora_into_transformer(
2545-
cls,
2532+
cls,
25462533
state_dict,
25472534
transformer,
25482535
adapter_name=None,
@@ -2739,7 +2726,7 @@ def load_lora_weights(
27392726
@classmethod
27402727
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->MochiTransformer3DModel
27412728
def load_lora_into_transformer(
2742-
cls,
2729+
cls,
27432730
state_dict,
27442731
transformer,
27452732
adapter_name=None,
@@ -2942,7 +2929,7 @@ def load_lora_weights(
29422929
@classmethod
29432930
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->LTXVideoTransformer3DModel
29442931
def load_lora_into_transformer(
2945-
cls,
2932+
cls,
29462933
state_dict,
29472934
transformer,
29482935
adapter_name=None,
@@ -3142,7 +3129,7 @@ def load_lora_weights(
31423129
@classmethod
31433130
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->SanaTransformer2DModel
31443131
def load_lora_into_transformer(
3145-
cls,
3132+
cls,
31463133
state_dict,
31473134
transformer,
31483135
adapter_name=None,
@@ -3345,7 +3332,7 @@ def load_lora_weights(
33453332
@classmethod
33463333
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->HunyuanVideoTransformer3DModel
33473334
def load_lora_into_transformer(
3348-
cls,
3335+
cls,
33493336
state_dict,
33503337
transformer,
33513338
adapter_name=None,
@@ -3549,7 +3536,7 @@ def load_lora_weights(
35493536
@classmethod
35503537
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->Lumina2Transformer2DModel
35513538
def load_lora_into_transformer(
3552-
cls,
3539+
cls,
35533540
state_dict,
35543541
transformer,
35553542
adapter_name=None,
@@ -4108,7 +4095,7 @@ def load_lora_weights(
41084095
@classmethod
41094096
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->WanTransformer3DModel
41104097
def load_lora_into_transformer(
4111-
cls,
4098+
cls,
41124099
state_dict,
41134100
transformer,
41144101
adapter_name=None,
@@ -4385,7 +4372,7 @@ def load_lora_weights(
43854372
@classmethod
43864373
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->SkyReelsV2Transformer3DModel
43874374
def load_lora_into_transformer(
4388-
cls,
4375+
cls,
43894376
state_dict,
43904377
transformer,
43914378
adapter_name=None,
@@ -4585,7 +4572,7 @@ def load_lora_weights(
45854572
@classmethod
45864573
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->CogView4Transformer2DModel
45874574
def load_lora_into_transformer(
4588-
cls,
4575+
cls,
45894576
state_dict,
45904577
transformer,
45914578
adapter_name=None,
@@ -4788,7 +4775,7 @@ def load_lora_weights(
47884775
@classmethod
47894776
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->HiDreamImageTransformer2DModel
47904777
def load_lora_into_transformer(
4791-
cls,
4778+
cls,
47924779
state_dict,
47934780
transformer,
47944781
adapter_name=None,
@@ -4994,7 +4981,7 @@ def load_lora_weights(
49944981
@classmethod
49954982
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->QwenImageTransformer2DModel
49964983
def load_lora_into_transformer(
4997-
cls,
4984+
cls,
49984985
state_dict,
49994986
transformer,
50004987
adapter_name=None,
@@ -5200,7 +5187,7 @@ def load_lora_weights(
52005187
@classmethod
52015188
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->ZImageTransformer2DModel
52025189
def load_lora_into_transformer(
5203-
cls,
5190+
cls,
52045191
state_dict,
52055192
transformer,
52065193
adapter_name=None,
@@ -5403,7 +5390,7 @@ def load_lora_weights(
54035390
@classmethod
54045391
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->CogView4Transformer2DModel
54055392
def load_lora_into_transformer(
5406-
cls,
5393+
cls,
54075394
state_dict,
54085395
transformer,
54095396
adapter_name=None,

src/diffusers/models/controlnets/multicontrolnet_union.py

Lines changed: 0 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -85,16 +85,6 @@ def forward(
8585

8686
# Copied from diffusers.models.controlnets.multicontrolnet.MultiControlNetModel.save_pretrained with ControlNet->ControlNetUnion
8787
def save_pretrained(
88-
self,
89-
save_directory: Union[str, os.PathLike],
90-
is_main_process: bool = True,
91-
save_function: Callable = None,
92-
safe_serialization: bool = True,
93-
variant: Optional[str] = None,
94-
):
95-
"""
96-
Save a model and its configuration file to a directory, so that it can be re-loaded using the
97-
`[`~models.controlnets.multicontrolnet.MultiControlNetUnionModel.from_pretrained`]` class method.
9888

9989
Arguments:
10090
save_directory (`str` or `os.PathLike`):
@@ -125,8 +115,6 @@ def save_pretrained(
125115
@classmethod
126116
# Copied from diffusers.models.controlnets.multicontrolnet.MultiControlNetModel.from_pretrained with ControlNet->ControlNetUnion
127117
def from_pretrained(cls, pretrained_model_path: Optional[Union[str, os.PathLike]], **kwargs):
128-
r"""
129-
Instantiate a pretrained MultiControlNetUnion model from multiple pre-trained controlnet models.
130118
131119
The model is set in evaluation mode by default using `model.eval()` (Dropout modules are deactivated). To train
132120
the model, you should first set it back in training mode with `model.train()`.

src/diffusers/models/transformers/auraflow_transformer_2d.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -367,9 +367,6 @@ def __init__(
367367

368368
# Copied from diffusers.models.unets.unet_2d_condition.UNet2DConditionModel.fuse_qkv_projections with FusedAttnProcessor2_0->FusedAuraFlowAttnProcessor2_0
369369
def fuse_qkv_projections(self):
370-
"""
371-
Enables fused QKV projections. For self-attention modules, all projection matrices (i.e., query, key, value)
372-
are fused. For cross-attention modules, key and value projection matrices are fused.
373370

374371
> [!WARNING] > This API is 🧪 experimental.
375372
"""

src/diffusers/models/transformers/cogvideox_transformer_3d.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -333,9 +333,6 @@ def __init__(
333333

334334
# Copied from diffusers.models.unets.unet_2d_condition.UNet2DConditionModel.fuse_qkv_projections with FusedAttnProcessor2_0->FusedCogVideoXAttnProcessor2_0
335335
def fuse_qkv_projections(self):
336-
"""
337-
Enables fused QKV projections. For self-attention modules, all projection matrices (i.e., query, key, value)
338-
are fused. For cross-attention modules, key and value projection matrices are fused.
339336

340337
> [!WARNING] > This API is 🧪 experimental.
341338
"""

src/diffusers/models/transformers/hunyuan_transformer_2d.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -320,9 +320,6 @@ def __init__(
320320

321321
# Copied from diffusers.models.unets.unet_2d_condition.UNet2DConditionModel.fuse_qkv_projections with FusedAttnProcessor2_0->FusedHunyuanAttnProcessor2_0
322322
def fuse_qkv_projections(self):
323-
"""
324-
Enables fused QKV projections. For self-attention modules, all projection matrices (i.e., query, key, value)
325-
are fused. For cross-attention modules, key and value projection matrices are fused.
326323

327324
> [!WARNING] > This API is 🧪 experimental.
328325
"""

src/diffusers/models/transformers/stable_audio_transformer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -276,7 +276,7 @@ def __init__(
276276

277277
# Copied from diffusers.models.transformers.hunyuan_transformer_2d.HunyuanDiT2DModel.set_default_attn_processor with Hunyuan->StableAudio
278278
def set_default_attn_processor(self):
279-
"""
279+
"""
280280
Disables custom attention processors and sets the default attention implementation.
281281
"""
282282
self.set_attn_processor(StableAudioAttnProcessor2_0())

src/diffusers/models/transformers/transformer_sd3.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -215,9 +215,6 @@ def fn_recursive_feed_forward(module: torch.nn.Module, chunk_size: int, dim: int
215215

216216
# Copied from diffusers.models.unets.unet_2d_condition.UNet2DConditionModel.fuse_qkv_projections with FusedAttnProcessor2_0->FusedJointAttnProcessor2_0
217217
def fuse_qkv_projections(self):
218-
"""
219-
Enables fused QKV projections. For self-attention modules, all projection matrices (i.e., query, key, value)
220-
are fused. For cross-attention modules, key and value projection matrices are fused.
221218

222219
> [!WARNING] > This API is 🧪 experimental.
223220
"""

src/diffusers/modular_pipelines/flux/before_denoise.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -273,7 +273,7 @@ def intermediate_outputs(self) -> List[OutputParam]:
273273
@staticmethod
274274
# Copied from diffusers.pipelines.stable_diffusion_3.pipeline_stable_diffusion_3_img2img.StableDiffusion3Img2ImgPipeline.get_timesteps with self.scheduler->scheduler
275275
def get_timesteps(scheduler, num_inference_steps, strength, device):
276-
# get the original timestep using init_timestep
276+
# get the original timestep using init_timestep
277277
init_timestep = min(num_inference_steps * strength, num_inference_steps)
278278

279279
t_start = int(max(num_inference_steps - init_timestep, 0))

src/diffusers/modular_pipelines/stable_diffusion_xl/before_denoise.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -443,7 +443,7 @@ def intermediate_outputs(self) -> List[str]:
443443
@staticmethod
444444
# Copied from diffusers.pipelines.stable_diffusion_xl.pipeline_stable_diffusion_xl_img2img.StableDiffusionXLImg2ImgPipeline.get_timesteps with self->components
445445
def get_timesteps(components, num_inference_steps, strength, device, denoising_start=None):
446-
# get the original timestep using init_timestep
446+
# get the original timestep using init_timestep
447447
if denoising_start is None:
448448
init_timestep = min(int(num_inference_steps * strength), num_inference_steps)
449449
t_start = max(num_inference_steps - init_timestep, 0)
@@ -1005,7 +1005,7 @@ def check_inputs(components, block_state):
10051005
@staticmethod
10061006
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline.prepare_latents with self->comp
10071007
def prepare_latents(comp, batch_size, num_channels_latents, height, width, dtype, device, generator, latents=None):
1008-
shape = (
1008+
shape = (
10091009
batch_size,
10101010
num_channels_latents,
10111011
int(height) // comp.vae_scale_factor,
@@ -1135,7 +1135,7 @@ def intermediate_outputs(self) -> List[OutputParam]:
11351135
@staticmethod
11361136
# Copied from diffusers.pipelines.stable_diffusion_xl.pipeline_stable_diffusion_xl_img2img.StableDiffusionXLImg2ImgPipeline._get_add_time_ids with self->components
11371137
def _get_add_time_ids(
1138-
components,
1138+
components,
11391139
original_size,
11401140
crops_coords_top_left,
11411141
target_size,
@@ -1346,7 +1346,7 @@ def intermediate_outputs(self) -> List[OutputParam]:
13461346
@staticmethod
13471347
# Copied from diffusers.pipelines.stable_diffusion_xl.pipeline_stable_diffusion_xl.StableDiffusionXLPipeline._get_add_time_ids with self->components
13481348
def _get_add_time_ids(
1349-
components, original_size, crops_coords_top_left, target_size, dtype, text_encoder_projection_dim=None
1349+
components, original_size, crops_coords_top_left, target_size, dtype, text_encoder_projection_dim=None
13501350
):
13511351
add_time_ids = list(original_size + crops_coords_top_left + target_size)
13521352

src/diffusers/modular_pipelines/stable_diffusion_xl/decoders.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -76,7 +76,7 @@ def intermediate_outputs(self) -> List[str]:
7676
@staticmethod
7777
# Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale.StableDiffusionUpscalePipeline.upcast_vae with self->components
7878
def upcast_vae(components):
79-
deprecate(
79+
deprecate(
8080
"upcast_vae",
8181
"1.0.0",
8282
"`upcast_vae` is deprecated. Please use `pipe.vae.to(torch.float32)`. For more details, please refer to: https://github.com/huggingface/diffusers/pull/12619#issue-3606633695.",

0 commit comments

Comments
 (0)