From 53ebca58ffaed88cdda13aaa989eb3bce9594de4 Mon Sep 17 00:00:00 2001 From: Ryan Dick Date: Wed, 28 Feb 2024 13:53:56 -0500 Subject: [PATCH] Rename ConditioningData to TextConditioningData. --- invokeai/app/invocations/latent.py | 6 +++--- .../backend/stable_diffusion/diffusers_pipeline.py | 11 +++++++---- .../stable_diffusion/diffusion/conditioning_data.py | 2 +- .../diffusion/shared_invokeai_diffusion.py | 10 +++++----- 4 files changed, 16 insertions(+), 13 deletions(-) diff --git a/invokeai/app/invocations/latent.py b/invokeai/app/invocations/latent.py index 7fc389e12f..9b6e14270c 100644 --- a/invokeai/app/invocations/latent.py +++ b/invokeai/app/invocations/latent.py @@ -43,9 +43,9 @@ from invokeai.backend.ip_adapter.ip_adapter import IPAdapter, IPAdapterPlus from invokeai.backend.model_management.models import ModelType, SilenceWarnings from invokeai.backend.stable_diffusion.diffusion.conditioning_data import ( BasicConditioningInfo, - ConditioningData, IPAdapterConditioningInfo, SDXLConditioningInfo, + TextConditioningData, ) from ...backend.model_management.lora import ModelPatcher @@ -359,7 +359,7 @@ class DenoiseLatentsInvocation(BaseInvocation): scheduler, unet, seed, - ) -> ConditioningData: + ) -> TextConditioningData: cond_text_embeddings, cond_text_embedding_masks = self._get_text_embeddings_and_masks( self.positive_conditioning, context, unet.device, unet.dtype ) @@ -367,7 +367,7 @@ class DenoiseLatentsInvocation(BaseInvocation): self.negative_conditioning, context, unet.device, unet.dtype ) - conditioning_data = ConditioningData( + conditioning_data = TextConditioningData( uncond_text_embeddings=uncond_text_embeddings, uncond_text_embedding_masks=uncond_text_embedding_masks, cond_text_embeddings=cond_text_embeddings, diff --git a/invokeai/backend/stable_diffusion/diffusers_pipeline.py b/invokeai/backend/stable_diffusion/diffusers_pipeline.py index 3e50530598..434f67fa13 100644 --- a/invokeai/backend/stable_diffusion/diffusers_pipeline.py +++ b/invokeai/backend/stable_diffusion/diffusers_pipeline.py @@ -24,7 +24,10 @@ from transformers import CLIPFeatureExtractor, CLIPTextModel, CLIPTokenizer from invokeai.app.services.config import InvokeAIAppConfig from invokeai.backend.ip_adapter.ip_adapter import IPAdapter from invokeai.backend.ip_adapter.unet_patcher import UNetPatcher -from invokeai.backend.stable_diffusion.diffusion.conditioning_data import ConditioningData, IPAdapterConditioningInfo +from invokeai.backend.stable_diffusion.diffusion.conditioning_data import ( + IPAdapterConditioningInfo, + TextConditioningData, +) from invokeai.backend.stable_diffusion.diffusion.regional_prompt_attention import apply_regional_prompt_attn from invokeai.backend.stable_diffusion.diffusion.shared_invokeai_diffusion import InvokeAIDiffuserComponent @@ -311,7 +314,7 @@ class StableDiffusionGeneratorPipeline(StableDiffusionPipeline): latents: torch.Tensor, num_inference_steps: int, scheduler_step_kwargs: dict[str, Any], - conditioning_data: ConditioningData, + conditioning_data: TextConditioningData, *, noise: Optional[torch.Tensor], timesteps: torch.Tensor, @@ -390,7 +393,7 @@ class StableDiffusionGeneratorPipeline(StableDiffusionPipeline): self, latents: torch.Tensor, timesteps, - conditioning_data: ConditioningData, + conditioning_data: TextConditioningData, scheduler_step_kwargs: dict[str, Any], *, additional_guidance: List[Callable] = None, @@ -487,7 +490,7 @@ class StableDiffusionGeneratorPipeline(StableDiffusionPipeline): self, t: torch.Tensor, latents: torch.Tensor, - conditioning_data: ConditioningData, + conditioning_data: TextConditioningData, step_index: int, total_step_count: int, scheduler_step_kwargs: dict[str, Any], diff --git a/invokeai/backend/stable_diffusion/diffusion/conditioning_data.py b/invokeai/backend/stable_diffusion/diffusion/conditioning_data.py index 98ad3dcfdc..9f00dc191a 100644 --- a/invokeai/backend/stable_diffusion/diffusion/conditioning_data.py +++ b/invokeai/backend/stable_diffusion/diffusion/conditioning_data.py @@ -54,7 +54,7 @@ class IPAdapterConditioningInfo: @dataclass -class ConditioningData: +class TextConditioningData: uncond_text_embeddings: Union[list[BasicConditioningInfo], list[SDXLConditioningInfo]] uncond_text_embedding_masks: list[Optional[torch.Tensor]] cond_text_embeddings: Union[list[BasicConditioningInfo], list[SDXLConditioningInfo]] diff --git a/invokeai/backend/stable_diffusion/diffusion/shared_invokeai_diffusion.py b/invokeai/backend/stable_diffusion/diffusion/shared_invokeai_diffusion.py index 5251197cb1..16c55fa57d 100644 --- a/invokeai/backend/stable_diffusion/diffusion/shared_invokeai_diffusion.py +++ b/invokeai/backend/stable_diffusion/diffusion/shared_invokeai_diffusion.py @@ -12,10 +12,10 @@ from typing_extensions import TypeAlias from invokeai.app.services.config import InvokeAIAppConfig from invokeai.backend.stable_diffusion.diffusion.conditioning_data import ( BasicConditioningInfo, - ConditioningData, ExtraConditioningInfo, IPAdapterConditioningInfo, SDXLConditioningInfo, + TextConditioningData, ) from invokeai.backend.stable_diffusion.diffusion.regional_prompt_attention import Range, RegionalPromptData @@ -230,7 +230,7 @@ class InvokeAIDiffuserComponent: timestep: torch.Tensor, step_index: int, total_step_count: int, - conditioning_data: ConditioningData, + conditioning_data: TextConditioningData, ): down_block_res_samples, mid_block_res_sample = None, None # HACK(ryan): Currently, we just take the first text embedding if there's more than one. We should probably @@ -329,7 +329,7 @@ class InvokeAIDiffuserComponent: self, sample: torch.Tensor, timestep: torch.Tensor, - conditioning_data: ConditioningData, + conditioning_data: TextConditioningData, ip_adapter_conditioning: Optional[list[IPAdapterConditioningInfo]], step_index: int, total_step_count: int, @@ -428,7 +428,7 @@ class InvokeAIDiffuserComponent: self, x, sigma, - conditioning_data: ConditioningData, + conditioning_data: TextConditioningData, ip_adapter_conditioning: Optional[list[IPAdapterConditioningInfo]], down_block_additional_residuals: Optional[torch.Tensor] = None, # for ControlNet mid_block_additional_residual: Optional[torch.Tensor] = None, # for ControlNet @@ -531,7 +531,7 @@ class InvokeAIDiffuserComponent: self, x: torch.Tensor, sigma, - conditioning_data: ConditioningData, + conditioning_data: TextConditioningData, ip_adapter_conditioning: Optional[list[IPAdapterConditioningInfo]], cross_attention_control_types_to_do: list[CrossAttentionType], down_block_additional_residuals: Optional[torch.Tensor] = None, # for ControlNet