From cb7855369dc84e9098a983e98bca8b7f29ad4132 Mon Sep 17 00:00:00 2001 From: CyberVy <72680847+CyberVy@users.noreply.github.com> Date: Sat, 8 Mar 2025 21:31:32 +0800 Subject: [PATCH 1/3] Update lora_pipeline.py --- src/diffusers/loaders/lora_pipeline.py | 44 +++++++++++++------------- 1 file changed, 22 insertions(+), 22 deletions(-) diff --git a/src/diffusers/loaders/lora_pipeline.py b/src/diffusers/loaders/lora_pipeline.py index c5cb27a35f3c..c22c5d8a4345 100644 --- a/src/diffusers/loaders/lora_pipeline.py +++ b/src/diffusers/loaders/lora_pipeline.py @@ -455,7 +455,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["unet", "text_encoder"], **kwargs): @@ -476,7 +476,7 @@ def unfuse_lora(self, components: List[str] = ["unet", "text_encoder"], **kwargs Whether to unfuse the text encoder LoRA parameters. If the text encoder wasn't monkey-patched with the LoRA parameters then it won't have any effect. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class StableDiffusionXLLoraLoaderMixin(LoraBaseMixin): @@ -904,7 +904,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["unet", "text_encoder", "text_encoder_2"], **kwargs): @@ -925,7 +925,7 @@ def unfuse_lora(self, components: List[str] = ["unet", "text_encoder", "text_enc Whether to unfuse the text encoder LoRA parameters. If the text encoder wasn't monkey-patched with the LoRA parameters then it won't have any effect. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class SD3LoraLoaderMixin(LoraBaseMixin): @@ -1312,7 +1312,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["transformer", "text_encoder", "text_encoder_2"], **kwargs): @@ -1333,7 +1333,7 @@ def unfuse_lora(self, components: List[str] = ["transformer", "text_encoder", "t Whether to unfuse the text encoder LoRA parameters. If the text encoder wasn't monkey-patched with the LoRA parameters then it won't have any effect. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class FluxLoraLoaderMixin(LoraBaseMixin): @@ -1847,7 +1847,7 @@ def fuse_lora( ) super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["transformer", "text_encoder"], **kwargs): @@ -1868,7 +1868,7 @@ def unfuse_lora(self, components: List[str] = ["transformer", "text_encoder"], * if hasattr(transformer, "_transformer_norm_layers") and transformer._transformer_norm_layers: transformer.load_state_dict(transformer._transformer_norm_layers, strict=False) - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) # We override this here account for `_transformer_norm_layers` and `_overwritten_params`. def unload_lora_weights(self, reset_to_overwritten_params=False): @@ -2570,7 +2570,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): @@ -2588,7 +2588,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class Mochi1LoraLoaderMixin(LoraBaseMixin): @@ -2873,7 +2873,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): @@ -2891,7 +2891,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class LTXVideoLoraLoaderMixin(LoraBaseMixin): @@ -3176,7 +3176,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): @@ -3194,7 +3194,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class SanaLoraLoaderMixin(LoraBaseMixin): @@ -3479,7 +3479,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): @@ -3497,7 +3497,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class HunyuanVideoLoraLoaderMixin(LoraBaseMixin): @@ -3785,7 +3785,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): @@ -3803,7 +3803,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class Lumina2LoraLoaderMixin(LoraBaseMixin): @@ -4093,7 +4093,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) # Copied from diffusers.loaders.lora_pipeline.SanaLoraLoaderMixin.unfuse_lora @@ -4112,7 +4112,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class WanLoraLoaderMixin(LoraBaseMixin): @@ -4398,7 +4398,7 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs ) # Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.unfuse_lora @@ -4417,7 +4417,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class LoraLoaderMixin(StableDiffusionLoraLoaderMixin): From d5f6fa529732fa8a5ab18d8582d4b4fb6634343f Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 10 Mar 2025 08:34:55 +0000 Subject: [PATCH 2/3] Apply style fixes --- src/diffusers/loaders/lora_pipeline.py | 66 +++++++++++++++++++++----- 1 file changed, 55 insertions(+), 11 deletions(-) diff --git a/src/diffusers/loaders/lora_pipeline.py b/src/diffusers/loaders/lora_pipeline.py index 937302556260..19097daee9f4 100644 --- a/src/diffusers/loaders/lora_pipeline.py +++ b/src/diffusers/loaders/lora_pipeline.py @@ -451,7 +451,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) def unfuse_lora(self, components: List[str] = ["unet", "text_encoder"], **kwargs): @@ -891,7 +895,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) def unfuse_lora(self, components: List[str] = ["unet", "text_encoder", "text_encoder_2"], **kwargs): @@ -1290,7 +1298,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.StableDiffusionXLLoraLoaderMixin.unfuse_lora with unet->transformer @@ -1828,7 +1840,11 @@ def fuse_lora( ) super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) def unfuse_lora(self, components: List[str] = ["transformer", "text_encoder"], **kwargs): @@ -2548,7 +2564,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): @@ -2852,7 +2872,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.unfuse_lora @@ -3157,7 +3181,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.unfuse_lora @@ -3462,7 +3490,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.unfuse_lora @@ -3770,7 +3802,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.unfuse_lora @@ -4079,7 +4115,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.SanaLoraLoaderMixin.unfuse_lora @@ -4384,7 +4424,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names, **kwargs + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.unfuse_lora From 051c742832680c1e0d67fd36098190a10a298a14 Mon Sep 17 00:00:00 2001 From: hlky Date: Tue, 11 Mar 2025 15:27:15 +0000 Subject: [PATCH 3/3] fix-copies --- src/diffusers/loaders/lora_pipeline.py | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/src/diffusers/loaders/lora_pipeline.py b/src/diffusers/loaders/lora_pipeline.py index 31acadc6ada9..467ae1ef125f 100644 --- a/src/diffusers/loaders/lora_pipeline.py +++ b/src/diffusers/loaders/lora_pipeline.py @@ -4733,7 +4733,11 @@ def fuse_lora( ``` """ super().fuse_lora( - components=components, lora_scale=lora_scale, safe_fusing=safe_fusing, adapter_names=adapter_names + components=components, + lora_scale=lora_scale, + safe_fusing=safe_fusing, + adapter_names=adapter_names, + **kwargs, ) # Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.unfuse_lora @@ -4752,7 +4756,7 @@ def unfuse_lora(self, components: List[str] = ["transformer"], **kwargs): components (`List[str]`): List of LoRA-injectable components to unfuse LoRA from. unfuse_transformer (`bool`, defaults to `True`): Whether to unfuse the UNet LoRA parameters. """ - super().unfuse_lora(components=components) + super().unfuse_lora(components=components, **kwargs) class LoraLoaderMixin(StableDiffusionLoraLoaderMixin):