| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | from typing import List, Optional, Union |
| |
|
| | import PIL.Image |
| | import torch |
| | from transformers import CLIPTokenizer |
| |
|
| | from diffusers.models import AutoencoderKL, ControlNetModel, UNet2DConditionModel |
| | from diffusers.schedulers import PNDMScheduler |
| | from diffusers.utils import ( |
| | logging, |
| | replace_example_docstring, |
| | ) |
| | from diffusers.utils.torch_utils import randn_tensor |
| | from diffusers.pipelines.blip_diffusion.blip_image_processing import BlipImageProcessor |
| | from diffusers.pipelines.blip_diffusion.modeling_blip2 import Blip2QFormerModel |
| | from diffusers.pipelines.blip_diffusion.modeling_ctx_clip import ContextCLIPTextModel |
| | from diffusers.pipelines.pipeline_utils import DiffusionPipeline, ImagePipelineOutput |
| | from diffusers.image_processor import PipelineImageInput, VaeImageProcessor |
| |
|
| | logger = logging.get_logger(__name__) |
| |
|
| | EXAMPLE_DOC_STRING = """ |
| | Examples: |
| | ```py |
| | >>> from diffusers.pipelines import BlipDiffusionControlNetPipeline |
| | >>> from diffusers.utils import load_image |
| | >>> from controlnet_aux import CannyDetector |
| | >>> import torch |
| | |
| | >>> blip_diffusion_pipe = BlipDiffusionControlNetPipeline.from_pretrained( |
| | ... "Salesforce/blipdiffusion-controlnet", torch_dtype=torch.float16 |
| | ... ).to("cuda") |
| | |
| | >>> style_subject = "flower" |
| | >>> tgt_subject = "teapot" |
| | >>> text_prompt = "on a marble table" |
| | |
| | >>> cldm_cond_image = load_image( |
| | ... "https://huggingface.co/datasets/ayushtues/blipdiffusion_images/resolve/main/kettle.jpg" |
| | ... ).resize((512, 512)) |
| | >>> canny = CannyDetector() |
| | >>> cldm_cond_image = canny(cldm_cond_image, 30, 70, output_type="pil") |
| | >>> style_image = load_image( |
| | ... "https://huggingface.co/datasets/ayushtues/blipdiffusion_images/resolve/main/flower.jpg" |
| | ... ) |
| | >>> guidance_scale = 7.5 |
| | >>> num_inference_steps = 50 |
| | >>> negative_prompt = "over-exposure, under-exposure, saturated, duplicate, out of frame, lowres, cropped, worst quality, low quality, jpeg artifacts, morbid, mutilated, out of frame, ugly, bad anatomy, bad proportions, deformed, blurry, duplicate" |
| | |
| | |
| | >>> output = blip_diffusion_pipe( |
| | ... text_prompt, |
| | ... style_image, |
| | ... cldm_cond_image, |
| | ... style_subject, |
| | ... tgt_subject, |
| | ... guidance_scale=guidance_scale, |
| | ... num_inference_steps=num_inference_steps, |
| | ... neg_prompt=negative_prompt, |
| | ... height=512, |
| | ... width=512, |
| | ... ).images |
| | >>> output[0].save("image.png") |
| | ``` |
| | """ |
| |
|
| |
|
| | class BlipDiffusionControlNetPipeline(DiffusionPipeline): |
| | """ |
| | Pipeline for Canny Edge based Controlled subject-driven generation using Blip Diffusion. |
| | |
| | This model inherits from [`DiffusionPipeline`]. Check the superclass documentation for the generic methods the |
| | library implements for all the pipelines (such as downloading or saving, running on a particular device, etc.) |
| | |
| | Args: |
| | tokenizer ([`CLIPTokenizer`]): |
| | Tokenizer for the text encoder |
| | text_encoder ([`ContextCLIPTextModel`]): |
| | Text encoder to encode the text prompt |
| | vae ([`AutoencoderKL`]): |
| | VAE model to map the latents to the image |
| | unet ([`UNet2DConditionModel`]): |
| | Conditional U-Net architecture to denoise the image embedding. |
| | scheduler ([`PNDMScheduler`]): |
| | A scheduler to be used in combination with `unet` to generate image latents. |
| | qformer ([`Blip2QFormerModel`]): |
| | QFormer model to get multi-modal embeddings from the text and image. |
| | controlnet ([`ControlNetModel`]): |
| | ControlNet model to get the conditioning image embedding. |
| | image_processor ([`BlipImageProcessor`]): |
| | Image Processor to preprocess and postprocess the image. |
| | ctx_begin_pos (int, `optional`, defaults to 2): |
| | Position of the context token in the text encoder. |
| | """ |
| |
|
| | model_cpu_offload_seq = "qformer->text_encoder->unet->vae" |
| |
|
| | def __init__( |
| | self, |
| | tokenizer: CLIPTokenizer, |
| | text_encoder: ContextCLIPTextModel, |
| | vae: AutoencoderKL, |
| | unet: UNet2DConditionModel, |
| | scheduler: PNDMScheduler, |
| | qformer: Blip2QFormerModel, |
| | controlnet: ControlNetModel, |
| | image_processor: BlipImageProcessor, |
| | ctx_begin_pos: int = 2, |
| | mean: List[float] = None, |
| | std: List[float] = None, |
| | ): |
| | super().__init__() |
| |
|
| | self.register_modules( |
| | tokenizer=tokenizer, |
| | text_encoder=text_encoder, |
| | vae=vae, |
| | unet=unet, |
| | scheduler=scheduler, |
| | qformer=qformer, |
| | controlnet=controlnet, |
| | image_processor=image_processor, |
| | ) |
| | |
| | self.vae_scale_factor = 2 ** (len(self.vae.config.block_out_channels) - 1) |
| | self.init_processor = VaeImageProcessor(vae_scale_factor=self.vae_scale_factor) |
| | self.mask_processor = VaeImageProcessor( |
| | vae_scale_factor=self.vae_scale_factor, do_normalize=False, do_binarize=True, do_convert_grayscale=True |
| | ) |
| | self.control_image_processor = VaeImageProcessor( |
| | vae_scale_factor=self.vae_scale_factor, do_convert_rgb=True, do_normalize=False |
| | ) |
| | self.register_to_config(ctx_begin_pos=ctx_begin_pos, mean=mean, std=std) |
| |
|
| | def get_query_embeddings(self, input_image, src_subject): |
| | return self.qformer(image_input=input_image, text_input=src_subject, return_dict=False) |
| |
|
| | |
| | def _build_prompt(self, prompts, tgt_subjects, prompt_strength=1.0, prompt_reps=20): |
| | rv = [] |
| | for prompt, tgt_subject in zip(prompts, tgt_subjects): |
| | prompt = f"a {tgt_subject} {prompt.strip()}" |
| | |
| | rv.append(", ".join([prompt] * int(prompt_strength * prompt_reps))) |
| |
|
| | return rv |
| |
|
| | |
| | def prepare_latents_old( |
| | self, |
| | batch_size, |
| | num_channels, |
| | height, |
| | width, |
| | dtype, |
| | device, |
| | generator, |
| | latents=None, |
| | image=None): |
| | shape = (batch_size, num_channels, height, width) |
| | if isinstance(generator, list) and len(generator) != batch_size: |
| | raise ValueError( |
| | f"You have passed a list of generators of length {len(generator)}, but requested an effective batch" |
| | f" size of {batch_size}. Make sure the batch size matches the length of the generators." |
| | ) |
| |
|
| | if latents is None: |
| | latents = randn_tensor(shape, generator=generator, device=device, dtype=dtype) |
| | else: |
| | latents = latents.to(device=device, dtype=dtype) |
| |
|
| | |
| | latents = latents * self.scheduler.init_noise_sigma |
| | return latents |
| |
|
| | |
| | def prepare_latents( |
| | self, |
| | batch_size, |
| | num_channels_latents, |
| | height, |
| | width, |
| | dtype, |
| | device, |
| | generator, |
| | latents=None, |
| | image=None, |
| | timestep=None, |
| | is_strength_max=True, |
| | return_noise=False, |
| | return_image_latents=False, |
| | ): |
| | shape = (batch_size, num_channels_latents, height // self.vae_scale_factor, width // self.vae_scale_factor) |
| | if isinstance(generator, list) and len(generator) != batch_size: |
| | raise ValueError( |
| | f"You have passed a list of generators of length {len(generator)}, but requested an effective batch" |
| | f" size of {batch_size}. Make sure the batch size matches the length of the generators." |
| | ) |
| |
|
| | if (image is None or timestep is None) and not is_strength_max: |
| | raise ValueError( |
| | "Since strength < 1. initial latents are to be initialised as a combination of Image + Noise." |
| | "However, either the image or the noise timestep has not been provided." |
| | ) |
| |
|
| | if return_image_latents or (latents is None and not is_strength_max): |
| | image = image.to(device=device, dtype=dtype) |
| |
|
| | if image.shape[1] == 4: |
| | image_latents = image |
| | else: |
| | image_latents = self._encode_vae_image(image=image, generator=generator) |
| | image_latents = image_latents.repeat(batch_size // image_latents.shape[0], 1, 1, 1) |
| |
|
| | if latents is None: |
| | noise = randn_tensor(shape, generator=generator, device=device, dtype=dtype) |
| | |
| | latents = noise if is_strength_max else self.scheduler.add_noise(image_latents, noise, timestep) |
| | |
| | latents = latents * self.scheduler.init_noise_sigma if is_strength_max else latents |
| | else: |
| | noise = latents.to(device) |
| | latents = noise * self.scheduler.init_noise_sigma |
| |
|
| | outputs = (latents,) |
| |
|
| | if return_noise: |
| | outputs += (noise,) |
| |
|
| | if return_image_latents: |
| | outputs += (image_latents,) |
| |
|
| | return outputs |
| |
|
| | def encode_prompt(self, query_embeds, prompt, device=None): |
| | device = device or self._execution_device |
| |
|
| | |
| | max_len = self.text_encoder.text_model.config.max_position_embeddings |
| | max_len -= self.qformer.config.num_query_tokens |
| |
|
| | tokenized_prompt = self.tokenizer( |
| | prompt, |
| | padding="max_length", |
| | truncation=True, |
| | max_length=max_len, |
| | return_tensors="pt", |
| | ).to(device) |
| |
|
| | batch_size = query_embeds.shape[0] |
| | ctx_begin_pos = [self.config.ctx_begin_pos] * batch_size |
| |
|
| | text_embeddings = self.text_encoder( |
| | input_ids=tokenized_prompt.input_ids, |
| | ctx_embeddings=query_embeds, |
| | ctx_begin_pos=ctx_begin_pos, |
| | )[0] |
| |
|
| | return text_embeddings |
| |
|
| | |
| | def get_timesteps(self, num_inference_steps, strength, device): |
| | |
| | init_timestep = min(int(num_inference_steps * strength), num_inference_steps) |
| |
|
| | t_start = max(num_inference_steps - init_timestep, 0) |
| | timesteps = self.scheduler.timesteps[t_start * self.scheduler.order :] |
| |
|
| | return timesteps, num_inference_steps - t_start |
| |
|
| | |
| | def _encode_vae_image(self, image: torch.Tensor, generator: torch.Generator): |
| | if isinstance(generator, list): |
| | image_latents = [ |
| | self.vae.encode(image[i : i + 1]).latent_dist.sample(generator=generator[i]) |
| | for i in range(image.shape[0]) |
| | ] |
| | image_latents = torch.cat(image_latents, dim=0) |
| | else: |
| | image_latents = self.vae.encode(image).latent_dist.sample(generator=generator) |
| |
|
| | image_latents = self.vae.config.scaling_factor * image_latents |
| |
|
| | return image_latents |
| | |
| | |
| | def prepare_mask_latents( |
| | self, mask, masked_image, batch_size, height, width, dtype, device, generator, do_classifier_free_guidance |
| | ): |
| | |
| | |
| | |
| | mask = torch.nn.functional.interpolate( |
| | mask, size=(height // self.vae_scale_factor, width // self.vae_scale_factor) |
| | ) |
| | mask = mask.to(device=device, dtype=dtype) |
| |
|
| | masked_image = masked_image.to(device=device, dtype=dtype) |
| |
|
| | if masked_image.shape[1] == 4: |
| | masked_image_latents = masked_image |
| | else: |
| | masked_image_latents = self._encode_vae_image(masked_image, generator=generator) |
| |
|
| | |
| | if mask.shape[0] < batch_size: |
| | if not batch_size % mask.shape[0] == 0: |
| | raise ValueError( |
| | "The passed mask and the required batch size don't match. Masks are supposed to be duplicated to" |
| | f" a total batch size of {batch_size}, but {mask.shape[0]} masks were passed. Make sure the number" |
| | " of masks that you pass is divisible by the total requested batch size." |
| | ) |
| | mask = mask.repeat(batch_size // mask.shape[0], 1, 1, 1) |
| | if masked_image_latents.shape[0] < batch_size: |
| | if not batch_size % masked_image_latents.shape[0] == 0: |
| | raise ValueError( |
| | "The passed images and the required batch size don't match. Images are supposed to be duplicated" |
| | f" to a total batch size of {batch_size}, but {masked_image_latents.shape[0]} images were passed." |
| | " Make sure the number of images that you pass is divisible by the total requested batch size." |
| | ) |
| | masked_image_latents = masked_image_latents.repeat(batch_size // masked_image_latents.shape[0], 1, 1, 1) |
| |
|
| | mask = torch.cat([mask] * 2) if do_classifier_free_guidance else mask |
| | masked_image_latents = ( |
| | torch.cat([masked_image_latents] * 2) if do_classifier_free_guidance else masked_image_latents |
| | ) |
| |
|
| | |
| | masked_image_latents = masked_image_latents.to(device=device, dtype=dtype) |
| | return mask, masked_image_latents |
| |
|
| | |
| | def prepare_control_image( |
| | self, |
| | image, |
| | width, |
| | height, |
| | batch_size, |
| | num_images_per_prompt, |
| | device, |
| | dtype, |
| | do_classifier_free_guidance=False, |
| | ): |
| | ''' |
| | image = self.control_image_processor.preprocess( |
| | image, |
| | height=height, |
| | width=width, |
| | #size={"width": width, "height": height}, |
| | do_rescale=True, |
| | do_center_crop=False, |
| | do_normalize=False, |
| | return_tensors="pt", |
| | )["pixel_values"].to(device) |
| | ''' |
| | image = self.control_image_processor.preprocess(image, height=height, width=width).to(dtype=torch.float32) |
| | image_batch_size = image.shape[0] |
| |
|
| | if image_batch_size == 1: |
| | repeat_by = batch_size |
| | else: |
| | |
| | repeat_by = num_images_per_prompt |
| |
|
| | image = image.repeat_interleave(repeat_by, dim=0) |
| |
|
| | image = image.to(device=device, dtype=dtype) |
| |
|
| | if do_classifier_free_guidance: |
| | image = torch.cat([image] * 2) |
| |
|
| | return image |
| |
|
| | @torch.no_grad() |
| | @replace_example_docstring(EXAMPLE_DOC_STRING) |
| | def __call__( |
| | self, |
| | prompt: List[str], |
| | reference_image: PIL.Image.Image, |
| | condtioning_image: PIL.Image.Image, |
| | source_subject_category: List[str], |
| | target_subject_category: List[str], |
| | image: PipelineImageInput = None, |
| | mask_image: PipelineImageInput = None, |
| | latents: Optional[torch.FloatTensor] = None, |
| | guidance_scale: float = 7.5, |
| | height: int = 512, |
| | width: int = 512, |
| | num_inference_steps: int = 50, |
| | generator: Optional[Union[torch.Generator, List[torch.Generator]]] = None, |
| | neg_prompt: Optional[str] = "", |
| | prompt_strength: float = 1.0, |
| | strength: float = 1.0, |
| | num_images_per_prompt: Optional[int] = 1, |
| | prompt_reps: int = 20, |
| | output_type: Optional[str] = "pil", |
| | return_dict: bool = True, |
| | ): |
| | """ |
| | Function invoked when calling the pipeline for generation. |
| | |
| | Args: |
| | prompt (`List[str]`): |
| | The prompt or prompts to guide the image generation. |
| | reference_image (`PIL.Image.Image`): |
| | The reference image to condition the generation on. |
| | condtioning_image (`PIL.Image.Image`): |
| | The conditioning canny edge image to condition the generation on. |
| | source_subject_category (`List[str]`): |
| | The source subject category. |
| | target_subject_category (`List[str]`): |
| | The target subject category. |
| | latents (`torch.FloatTensor`, *optional*): |
| | Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image |
| | generation. Can be used to tweak the same generation with different prompts. If not provided, a latents |
| | tensor will ge generated by random sampling. |
| | guidance_scale (`float`, *optional*, defaults to 7.5): |
| | Guidance scale as defined in [Classifier-Free Diffusion Guidance](https://arxiv.org/abs/2207.12598). |
| | `guidance_scale` is defined as `w` of equation 2. of [Imagen |
| | Paper](https://arxiv.org/pdf/2205.11487.pdf). Guidance scale is enabled by setting `guidance_scale > |
| | 1`. Higher guidance scale encourages to generate images that are closely linked to the text `prompt`, |
| | usually at the expense of lower image quality. |
| | height (`int`, *optional*, defaults to 512): |
| | The height of the generated image. |
| | width (`int`, *optional*, defaults to 512): |
| | The width of the generated image. |
| | seed (`int`, *optional*, defaults to 42): |
| | The seed to use for random generation. |
| | num_inference_steps (`int`, *optional*, defaults to 50): |
| | The number of denoising steps. More denoising steps usually lead to a higher quality image at the |
| | expense of slower inference. |
| | generator (`torch.Generator` or `List[torch.Generator]`, *optional*): |
| | One or a list of [torch generator(s)](https://pytorch.org/docs/stable/generated/torch.Generator.html) |
| | to make generation deterministic. |
| | neg_prompt (`str`, *optional*, defaults to ""): |
| | The prompt or prompts not to guide the image generation. Ignored when not using guidance (i.e., ignored |
| | if `guidance_scale` is less than `1`). |
| | prompt_strength (`float`, *optional*, defaults to 1.0): |
| | The strength of the prompt. Specifies the number of times the prompt is repeated along with prompt_reps |
| | to amplify the prompt. |
| | prompt_reps (`int`, *optional*, defaults to 20): |
| | The number of times the prompt is repeated along with prompt_strength to amplify the prompt. |
| | Examples: |
| | |
| | Returns: |
| | [`~pipelines.ImagePipelineOutput`] or `tuple` |
| | """ |
| | device = self._execution_device |
| |
|
| | reference_image = self.image_processor.preprocess( |
| | reference_image, image_mean=self.config.mean, image_std=self.config.std, return_tensors="pt" |
| | )["pixel_values"] |
| | reference_image = reference_image.to(device) |
| |
|
| | if isinstance(prompt, str): |
| | prompt = [prompt] |
| | if isinstance(source_subject_category, str): |
| | source_subject_category = [source_subject_category] |
| | if isinstance(target_subject_category, str): |
| | target_subject_category = [target_subject_category] |
| |
|
| | batch_size = len(prompt) |
| |
|
| | prompt = self._build_prompt( |
| | prompts=prompt, |
| | tgt_subjects=target_subject_category, |
| | prompt_strength=prompt_strength, |
| | prompt_reps=prompt_reps, |
| | ) |
| | query_embeds = self.get_query_embeddings(reference_image, source_subject_category) |
| | text_embeddings = self.encode_prompt(query_embeds, prompt, device) |
| | |
| | do_classifier_free_guidance = guidance_scale > 1.0 |
| | if do_classifier_free_guidance: |
| | max_length = self.text_encoder.text_model.config.max_position_embeddings |
| |
|
| | uncond_input = self.tokenizer( |
| | [neg_prompt] * batch_size, |
| | padding="max_length", |
| | max_length=max_length, |
| | return_tensors="pt", |
| | ) |
| | uncond_embeddings = self.text_encoder( |
| | input_ids=uncond_input.input_ids.to(device), |
| | ctx_embeddings=None, |
| | )[0] |
| | |
| | |
| | |
| | text_embeddings = torch.cat([uncond_embeddings, text_embeddings]) |
| | |
| | |
| | cond_image = self.prepare_control_image( |
| | image=condtioning_image, |
| | width=width, |
| | height=height, |
| | batch_size=batch_size, |
| | num_images_per_prompt=1, |
| | device=device, |
| | dtype=self.controlnet.dtype, |
| | do_classifier_free_guidance=do_classifier_free_guidance, |
| | ) |
| |
|
| | |
| | |
| | init_image = self.init_processor.preprocess(image, height=height, width=width) |
| | init_image = init_image.to(dtype=torch.float32) |
| |
|
| | mask = self.mask_processor.preprocess(mask_image, height=height, width=width) |
| |
|
| | masked_image = init_image * (mask < 0.5) |
| | _, _, height, width = init_image.shape |
| |
|
| | |
| | extra_set_kwargs = {} |
| | self.scheduler.set_timesteps(num_inference_steps, **extra_set_kwargs) |
| | timesteps, num_inference_steps = self.get_timesteps( |
| | num_inference_steps=num_inference_steps, strength=strength, device=device |
| | ) |
| | |
| | latent_timestep = timesteps[:1].repeat(batch_size * num_images_per_prompt) |
| | |
| | is_strength_max = strength == 1.0 |
| |
|
| | |
| | num_channels_latents = self.vae.config.latent_channels |
| | num_channels_unet = self.unet.config.in_channels |
| | return_image_latents = num_channels_unet == 4 |
| |
|
| | |
| | scale_down_factor = 2 ** (len(self.unet.config.block_out_channels) - 1) |
| | ''' |
| | latents = self.prepare_latents( |
| | batch_size=batch_size, |
| | num_channels=self.unet.config.in_channels, |
| | height=height // scale_down_factor, |
| | width=width // scale_down_factor, |
| | generator=generator, |
| | latents=latents, |
| | dtype=self.unet.dtype, |
| | device=device, |
| | image=init_image, |
| | ) |
| | ''' |
| | latents_outputs = self.prepare_latents( |
| | batch_size, |
| | num_channels_latents, |
| | height, |
| | width, |
| | text_embeddings.dtype, |
| | device, |
| | generator, |
| | latents, |
| | image=init_image, |
| | timestep=latent_timestep, |
| | is_strength_max=is_strength_max, |
| | return_noise=True, |
| | return_image_latents=return_image_latents, |
| | ) |
| |
|
| | if return_image_latents: |
| | latents, noise, image_latents = latents_outputs |
| | else: |
| | latents, noise = latents_outputs |
| |
|
| | |
| | mask, masked_image_latents = self.prepare_mask_latents( |
| | mask, |
| | masked_image, |
| | batch_size, |
| | height, |
| | width, |
| | text_embeddings.dtype, |
| | device, |
| | generator, |
| | do_classifier_free_guidance, |
| | ) |
| |
|
| | |
| | for i, t in enumerate(self.progress_bar(self.scheduler.timesteps)): |
| | |
| | do_classifier_free_guidance = guidance_scale > 1.0 |
| |
|
| | latent_model_input = torch.cat([latents] * 2) if do_classifier_free_guidance else latents |
| | latent_model_input = self.scheduler.scale_model_input(latent_model_input, t) |
| |
|
| | down_block_res_samples, mid_block_res_sample = self.controlnet( |
| | latent_model_input, |
| | t, |
| | encoder_hidden_states=text_embeddings, |
| | controlnet_cond=cond_image, |
| | return_dict=False, |
| | ) |
| |
|
| | noise_pred = self.unet( |
| | latent_model_input, |
| | timestep=t, |
| | encoder_hidden_states=text_embeddings, |
| | down_block_additional_residuals=down_block_res_samples, |
| | mid_block_additional_residual=mid_block_res_sample, |
| | )["sample"] |
| |
|
| | |
| | if do_classifier_free_guidance: |
| | noise_pred_uncond, noise_pred_text = noise_pred.chunk(2) |
| | noise_pred = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) |
| |
|
| | latents = self.scheduler.step( |
| | noise_pred, |
| | t, |
| | latents, |
| | )["prev_sample"] |
| |
|
| | if num_channels_unet == 4: |
| | init_latents_proper = image_latents |
| | if do_classifier_free_guidance: |
| | init_mask, _ = mask.chunk(2) |
| | else: |
| | init_mask = mask |
| |
|
| | if i < len(timesteps) - 1: |
| | noise_timestep = timesteps[i + 1] |
| | init_latents_proper = self.scheduler.add_noise( |
| | init_latents_proper, noise, torch.tensor([noise_timestep]) |
| | ) |
| |
|
| | latents = (1 - init_mask) * init_latents_proper + init_mask * latents |
| |
|
| |
|
| | image = self.vae.decode(latents / self.vae.config.scaling_factor, return_dict=False)[0] |
| | image = self.image_processor.postprocess(image, output_type=output_type) |
| |
|
| | |
| | self.maybe_free_model_hooks() |
| |
|
| | if not return_dict: |
| | return (image,) |
| |
|
| | return ImagePipelineOutput(images=image) |
| |
|