mmagic.models.editors.disco_diffusion.guider
¶
Module Contents¶
Classes¶
Each iteration, the AI cuts the image into smaller pieces known as cuts. |
|
Dango233(https://github.com/Dango233)'s version of MakeCutouts. |
|
Disco-Diffusion uses text and images to guide image generation. We will |
Functions¶
|
Sinc function. |
|
Lanczos filter's reconstruction kernel L(x). |
|
_summary_ |
|
Lanczos resampling image. |
|
range loss. |
|
spherical distance loss. |
|
Parse prompt, return text and text weight. |
|
Split prompts to a list of prompts. |
Attributes¶
- mmagic.models.editors.disco_diffusion.guider.sinc(x)[source]¶
Sinc function. If x equal to 0,
sinc(x) = 1
- else:
sinc(x) = sin(x)/ x
- Parameters
x (torch.Tensor) – Input Tensor
- Returns
Function output.
- Return type
torch.Tensor
- mmagic.models.editors.disco_diffusion.guider.lanczos(x, a)[source]¶
Lanczos filter’s reconstruction kernel L(x).
- mmagic.models.editors.disco_diffusion.guider.ramp(ratio, width)[source]¶
_summary_
- Parameters
ratio (_type_) – _description_
width (_type_) – _description_
- Returns
_description_
- Return type
_type_
- mmagic.models.editors.disco_diffusion.guider.resample(input, size, align_corners=True)[source]¶
Lanczos resampling image.
- Parameters
input (torch.Tensor) – Input image tensor.
size (Tuple[int, int]) – Output image size.
align_corners (bool) – align_corners argument of F.interpolate. Defaults to True.
- Returns
Resampling results.
- Return type
torch.Tensor
- mmagic.models.editors.disco_diffusion.guider.spherical_dist_loss(x, y)[source]¶
spherical distance loss.
- class mmagic.models.editors.disco_diffusion.guider.MakeCutouts(cut_size, cutn)[source]¶
Bases:
torch.nn.Module
Each iteration, the AI cuts the image into smaller pieces known as cuts.
, and compares each cut to the prompt to decide how to guide the next diffusion step. This classes will randomly cut patches and perform image augmentation to these patches.
- Parameters
cut_size (int) – Size of the patches.
cutn (int) – Number of patches to cut.
- class mmagic.models.editors.disco_diffusion.guider.MakeCutoutsDango(cut_size, Overview=4, InnerCrop=0, IC_Size_Pow=0.5, IC_Grey_P=0.2)[source]¶
Bases:
torch.nn.Module
Dango233(https://github.com/Dango233)’s version of MakeCutouts.
The improvement compared to
MakeCutouts
is that it use partial greyscale augmentation to capture structure, and partial rotation augmentation to capture whole frames.- Parameters
cut_size (int) – Size of the patches.
Overview (int) – The total number of overview cuts.
details (In) – Overview=1, Add whole frame; Overview=2, Add grayscaled frame; Overview=3, Add horizontal flip frame; Overview=4, Add grayscaled horizontal flip frame; Overview>4, Repeat add frame Overview times. Defaults to 4.
- :paramOverview=1, Add whole frame;
Overview=2, Add grayscaled frame; Overview=3, Add horizontal flip frame; Overview=4, Add grayscaled horizontal flip frame; Overview>4, Repeat add frame Overview times. Defaults to 4.
- Parameters
InnerCrop (int) – The total number of inner cuts. Defaults to 0.
IC_Size_Pow (float) – This sets the size of the border used for inner cuts. High values have larger borders, and therefore the cuts themselves will be smaller and provide finer details. Defaults to 0.5.
IC_Grey_P (float) – The portion of the inner cuts can be set to be grayscale instead of color. This may help with improved definition of shapes and edges, especially in the early diffusion steps where the image structure is being defined. Defaults to 0.2.
- mmagic.models.editors.disco_diffusion.guider.parse_prompt(prompt)[source]¶
Parse prompt, return text and text weight.
- mmagic.models.editors.disco_diffusion.guider.split_prompts(prompts, max_frames=1)[source]¶
Split prompts to a list of prompts.
- class mmagic.models.editors.disco_diffusion.guider.ImageTextGuider(clip_models)[source]¶
Bases:
torch.nn.Module
Disco-Diffusion uses text and images to guide image generation. We will use the clip models to extract text and image features as prompts, and then during the iteration, the features of the image patches are computed, and the similarity loss between the prompts features and the generated features is computed. Other losses also include RGB Range loss, total variation loss. Using these losses we can guide the image generation towards the desired target.
- Parameters
clip_models (List[Dict]) – List of clip model settings.
- property device[source]¶
Get current device of the model.
- Returns
The current device of the model.
- Return type
torch.device
- compute_prompt_stats(text_prompts=[], image_prompt=None, fuzzy_prompt=False, rand_mag=0.05)[source]¶
Compute prompts statistics.
- Parameters
text_prompts (list) – Text prompts. Defaults to [].
image_prompt (list) – Image prompts. Defaults to None.
fuzzy_prompt (bool, optional) – Controls whether to add multiple noisy prompts to the prompt losses. If True, can increase variability of image output. Defaults to False.
rand_mag (float, optional) – Controls the magnitude of the random noise added by fuzzy_prompt. Defaults to 0.05.
- cond_fn(model, diffusion_scheduler, x, t, beta_prod_t, model_stats, secondary_model=None, init_image=None, clamp_grad=True, clamp_max=0.05, clip_guidance_scale=5000, init_scale=1000, tv_scale=0.0, sat_scale=0.0, range_scale=150, cut_overview=[12] * 400 + [4] * 600, cut_innercut=[4] * 400 + [12] * 600, cut_ic_pow=[1] * 1000, cut_icgray_p=[0.2] * 400 + [0] * 600, cutn_batches=4)[source]¶
Clip guidance function.
- Parameters
model (nn.Module) – _description_
diffusion_scheduler (object) – _description_
x (torch.Tensor) – _description_
t (int) – _description_
beta_prod_t (torch.Tensor) – _description_
model_stats (List[torch.Tensor]) – _description_
secondary_model (nn.Module) – A smaller secondary diffusion model trained by Katherine Crowson to remove noise from intermediate timesteps to prepare them for CLIP. Ref: https://twitter.com/rivershavewings/status/1462859669454536711 # noqa Defaults to None.
init_image (torch.Tensor) – Initial image for denoising. Defaults to None.
clamp_grad (bool, optional) – Whether clamp gradient. Defaults to True.
clamp_max (float, optional) – Clamp max values. Defaults to 0.05.
clip_guidance_scale (int, optional) – The scale of influence of clip guidance on image generation. Defaults to 5000.