Extracted Text


pdf

Under review as a conference paper at ICLR 2024
SDXL:IMPROVINGLATENTDIFFUSIONMODELS FOR
HIGH-RESOLUTIONIMAGESYNTHESIS
Anonymous authors
Paper under double-blind review
ABSTRACT
We presentStable Diffusion XL(SDXL), a latent diffusion model for text-to-image
synthesis. Compared to previous versions ofStable Diffusion,SDXLleverages
a three times larger UNet backbone, achieved by significantly increasing the
number of attention blocks and including a second text encoder. Further, we design
multiple novel conditioning schemes and trainSDXLon multiple aspect ratios.
To ensure highest quality results, we also introduce arefinement modelwhich is
used to improve the visual fidelity of samples generated bySDXLusing a post-hoc
image-to-imagetechnique. We demonstrate thatSDXLimproves dramatically over
previous versions ofStable Diffusionand achieves results competitive with those
of black-box state-of-the-art image generators such as Midjourney (Holz, 2023).
1 INTRODUCTION
The last year has brought enormous leaps in deep generative modeling across various data domains,
such as natural language (Touvron et al., 2023), audio (Huang et al., 2023), and visual media (Rom-
bach et al., 2021; Ramesh et al., 2022; Saharia et al., 2022; Singer et al., 2022; Ho et al., 2022;
Blattmann et al., 2023; Esser et al., 2023). In this report, we focus on the latter and unveilSDXL,
a drastically improved version ofStable Diffusion.Stable Diffusionis a latent text-to-image dif-
fusion model (DM) which serves as the foundation for an array of recent advancements in, e.g.,
3D classification (Shen et al., 2023), controllable image editing (Zhang & Agrawala, 2023), image
1

Under review as a conference paper at ICLR 2024
personalization (Gal et al., 2022), synthetic data augmentation (Stöckl, 2022), graphical user interface
prototyping (Wei et al., 2023), etc. Remarkably, the scope of applications has been extraordinarily
extensive, encompassing fields as diverse as music generation (Forsgren & Martiros, 2022) and
reconstructing images from fMRI brain scans (Takagi & Nishimoto, 2023).
User studies demonstrate thatSDXLconsistently surpasses all previous versions ofStable Diffusion
by a significant margin (see Fig. 1). In this report, we present the design choices which lead to this
boost in performance encompassingi)a 3×larger UNet-backbone compared to previousStable
Diffusionmodels (Sec. 2.1),ii)two simple yet effective additional conditioning techniques (Sec. 2.2)
which do not require any form of additional supervision, andiii)a separate diffusion-based refinement
model which applies a noising-denoising process (Meng et al., 2021) to the latents produced bySDXL
to improve the visual quality of its samples (Sec. 2.5).
A major concern in the field of visual media creation is that while black-box-models are often
recognized as state-of-the-art, the opacity of their architecture prevents faithfully assessing and
validating their performance. This lack of transparency hampers reproducibility, stifles innovation,
and prevents the community from building upon these models to further the progress of science and
art. Moreover, these closed-source strategies make it challenging to assess the biases and limitations
of these models in an impartial and objective way, which is crucial for their responsible and ethical
deployment. WithSDXLwe are releasing anopenmodel that achieves competitive performance with
black-box image generation models (see Fig. 11 & Fig. 12).
2 IMPROVINGStable Diffusion
In this section we present our improvements for theStable Diffusionarchitecture. These are modular,
and can be used individually or together to extend any model. Although the following strategies are
implemented as extensions to latent diffusion models (LDMs) (Rombach et al., 2021), most of them
are also applicable to their pixel-space counterparts.
Figure 1:Left:Comparing user preferences betweenSDXLandStable Diffusion1.5 & 2.1. WhileSDXLalready
clearly outperformsStable Diffusion1.5 & 2.1, adding the additional refinement stage boosts performance.Right:
Visualization of the two-stage pipeline: We generate initial latents of size128×128 usingSDXL. Afterwards,
we utilize a specialized high-resolutionrefinement modeland apply SDEdit (Meng et al., 2021) on the latents
generated in the first step, using the same prompt.SDXLand the refinement model use the same autoencoder.
2.1 ARCHITECTURE& SCALE
Starting with the seminal works Ho et al. (2020) and Song et al. (2020b), which demonstrated that
DMs are powerful generative models for image synthesis, the convolutional UNet (Ronneberger
et al., 2015) architecture has been the dominant architecture for diffusion-based image synthesis.
However, with the development of foundational DMs (Saharia et al., 2022; Ramesh et al., 2022;
Rombach et al., 2021), the underlying architecture has constantly evolved: from adding self-attention
and improved upscaling layers (Dhariwal & Nichol, 2021), over cross-attention for text-to-image
synthesis (Rombach et al., 2021), to pure transformer-based architectures (Peebles & Xie, 2022).
We follow this trend and, following Hoogeboom et al. (2023), shift the bulk of the transformer
computation to lower-level features in the UNet. In particular, and in contrast to the originalStable
2

Under review as a conference paper at ICLR 2024
Table 1: Comparison ofSDXLand olderStable Diffusionmodels.
Model SDXL SD 1.4/1.5 SD 2.0/2.1
# of UNet params 2.6B 860M 865M
Transformer blocks [0, 2, 10] [1, 1, 1, 1] [1, 1, 1, 1]
Channel mult. [1, 2, 4] [1, 2, 4, 4] [1, 2, 4, 4]
Text encoder CLIP ViT-L & OpenCLIP ViT-bigG CLIP ViT-L OpenCLIP ViT-H
Context dim. 2048 768 1024
Pooled text emb. OpenCLIP ViT-bigG N/A N/A
Diffusionarchitecture, we use a heterogeneous distribution of transformer blocks within the UNet:
For efficiency reasons, we omit the transformer block at the highest feature level, use 2 and 10
blocks at the lower levels, and remove the lowest level (8×downsampling) in the UNet altogether
— see Tab. 1 for a comparison between the architectures ofStable Diffusion1.x & 2.x andSDXL.
We opt for a more powerful pre-trained text encoder that we use for text conditioning. Specifically,
we use OpenCLIP ViT-bigG (Ilharco et al., 2021) in combination with CLIP ViT-L (Radford et al.,
2021), where we concatenate the penultimate text encoder outputs along the channel-axis (Balaji
et al., 2022). Besides using cross-attention layers to condition the model on the text-input, we follow
Nichol et al. (2021) and additionally condition the model on the pooled text embedding from the
OpenCLIP model. These changes result in a model size of 2.6B parameters in the UNet, see Tab. 1.
The text encoders have a total size of 817M parameters.
2.2 MICRO-CONDITIONING
Figure 2: Height-vs-Width distribution of our
pre-training dataset. Without the proposed size-
conditioning, 39% of the data would be discarded due
to edge lengths smaller than 256 pixels as visualized
by the dashed black lines. Color intensity in each visu-
alized cell is proportional to the number of samples.
Conditioning the Model on Image SizeA no-
torious shortcoming of the LDM paradigm (Rom-
bach et al., 2021) is the fact that training a model
requires aminimal image size, due to its two-
stage architecture. The two main approaches to
tackle this problem are either to discard all train-
ing images below a certain minimal resolution
(for example,Stable Diffusion1.4/1.5 discarded
all images with any size below 512 pixels), or,
alternatively, upscale images that are too small.
However, depending on the desired image res-
olution, the former method can lead to signifi-
cant portions of the training data being discarded,
what will likely lead to a loss in performance
and hurt generalization. We visualize such ef-
fects in Fig. 2 for the dataset on whichSDXL
was pretrained. For this particular choice of data,
discarding all samples below our pretraining res-
olution of256
2pixels would lead to a significant 39% of discarded data. The second method, on
the other hand, usually introduces upscaling artifacts which may leak into the final model outputs,
causing, for example, blurry samples.
Instead, we propose to condition the UNet model on the original image resolution, which is trivially
available during training. In particular, we provide the original (i.e., before any rescaling) height
and width of the images as an additional conditioning to the modelc
size
= (horiginal, woriginal) .
Each component is independently embedded using a Fourier feature encoding, and these encodings
are concatenated into a single vector that we feed into the model by adding it to the timestep
embedding (Dhariwal & Nichol, 2021).
At inference time, a user can then set the desiredapparent resolutionof the image via thissize-
conditioning. Evidently (see Fig. 3), the model has learned to associate the conditioningcsizewith
resolution-dependent image features, which can be leveraged to modify the appearance of an output
corresponding to a given prompt. Note that for the visualization shown in Fig. 3, we visualize samples
generated by the512×512 model (see Sec. 2.5 for details), since the effects of the size conditioning
3

Under review as a conference paper at ICLR 2024
c
size
= (64,64) c
size
= (128,128), c
size
= (256,256), c
size
= (512,512),
“A robot painted as graffiti on a brick wall. a sidewalk is in front of the wall, and grass is growing out of cracks in the concrete.”“Panda mad scientist mixing sparkling chemicals, artstation.”
Figure 3: The effects of varying the size-conditioning: We show draw 4 samples with the same random seed from
SDXLand vary the size-conditioning as depicted above each column. The image quality clearly increases when
conditioning on larger image sizes. Samples from the512
2model, see Sec. 2.5. Note: For this visualization, we
use the512×512 pixel base model (see Sec. 2.5), since the effect of size conditioning is more clearly visible
before1024×1024finetuning. Best viewed zoomed in.
are less clearly visible after the subsequent multi-aspect (ratio) finetuning which we use for our final
SDXLmodel.
Table 2: Conditioning on the original spatial
size of the training examples improves perfor-
mance on class-conditional ImageNet Deng
et al. (2009) on512
2
resolution.
model FID-5k ↓IS-5k↑
CIN-512-only43.84 110.64
CIN-nocond 39.76 211.50
CIN-size-cond36.53 215.34
We quantitatively assess the effects of this simple but
effective conditioning technique by training and evaluating
three LDMs on class conditional ImageNet (Deng et al.,
2009) at spatial size512
2: For the first model (CIN-512-
only) we discard all training examples with at least one
edge smaller than512pixels what results in a train dataset
of only 70k images. ForCIN-nocondwe use all training
examples but without size conditioning. This additional
conditioning is only used forCIN-size-cond. After training
we generate 5k samples with 50 DDIM steps (Song et al.,
2020a) and (classifier-free) guidance scale of 5 (Ho & Salimans, 2022) for every model and compute
IS Salimans et al. (2016) and FID Heusel et al. (2017) (against the full validation set). ForCIN-
size-condwe generate samples always conditioned onc
size
= (512,512) . Tab. 2 summarizes the
results and verifies thatCIN-size-condimproves upon the baseline models in both metrics. We
attribute the degraded performance ofCIN-512-onlyto bad generalization due to overfitting on the
small training dataset while the effects of a mode of blurry samples in the sample distribution of
CIN-nocondresult in a reduced FID score. Note that, although we find these classical quantitative
scores not to be suitable for evaluating the performance of foundational (text-to-image) DMs Saharia
et al. (2022); Ramesh et al. (2022); Rombach et al. (2021) (see App. E), they remain reasonable
metrics on ImageNet as the neural backbones of FID and IS have been trained on ImageNet itself.
Conditioning the Model on Cropping ParametersThe first two rows of Fig. 4 illustrate a typical
failure mode of previousSDmodels: Synthesized objects can be cropped, such as the cut-off head
of the cat in the left examples forSD1-5 andSD2-1. An intuitive explanation for this behavior is
the use ofrandom croppingduring training of the model: As collating a batch in DL frameworks
such as PyTorch (Paszke et al., 2019) requires tensors of the same size, a typical processing pipeline
is to (i) resize an image such that the shortest size matches the desired target size, followed by (ii)
4

Under review as a conference paper at ICLR 2024
“A propaganda poster depicting a cat dressed as french
emperor napoleon holding a piece of cheese.”
“a close-up of a fire spitting dragon,
cinematic shot.”
SD1-5SD2-1SDXL
Figure 4: Comparison of the output ofSDXLwith previous versions ofStable Diffusion. For each prompt, we
show 3 random samples of the respective model for 50 steps of the DDIM sampler Song et al. (2020a) and
cfg-scale8.0Ho & Salimans (2022). Additional samples in Fig. 15.
randomly cropping the image along the longer axis. While random cropping is a natural form of data
augmentation, it can leak into the generated samples, causing the malicious effects shown above.
To fix this problem, we propose another simple yet effective conditioning method: During dataloading,
we uniformly sample crop coordinatesctopandcleft(integers specifying the amount of pixels cropped
from the top-left corner along the height and width axes, respectively) and feed them into the model
as conditioning parameters via Fourier feature embeddings, similar to the size conditioning described
above. The concatenated embeddingccropis then used as an additional conditioning parameter.
We emphasize that this technique is not limited to LDMs and could be used for any DM. Note that
crop- and size-conditioning can be readily combined. In such a case, we concatenate the feature
embedding along the channel dimension, before adding it to the timestep embedding in the UNet.
Alg. 1 illustrates how we sampleccropandc
sizeduring training if such a combination is applied.
Algorithm 1Size- and crop-micro-conditioning
Require:Training dataset of imagesD
Require:Target image size for trainings= (htgt, wtgt)
Require:Resizing functionR
Require:cropping function functionC
Require:Model train stepT
converged←False
whilenot convergeddo
x∼D
w
original←width(x)
h
original←height(x)
c
size
←(h
original, w
original)
x←R(x,s) ▷resize smaller image size to target sizes
ifh
original≤w
originalthen
c
left∼U(0,width(x)−sw) ▷samplec
left
ctop= 0
else ifh
original> w
originalthen
ctop∼U(0,height(x)−sh) ▷samplectop
c
left= 0
end if
ccrop←

ctop, c
left
·
x←C(x,s,ccrop) ▷crop image to sizes
converged←T(x,c
size
,ccrop) ▷train model
end while
Given that in our experience large scale
datasets are, on average, object-centric, we set
(ctop, cleft) = (0,0) during inference and thereby
obtain object-centered samples from the trained
model.
See Fig. 5 for an illustration: By tuning
(ctop, cleft) , we can successfullysimulatethe
amount of cropping during inference. This is
a form ofconditioning-augmentation, and has
been used in various forms with AR (Jun et al.,
2020) models, and recently with diffusion mod-
els (Karras et al., 2022).
While other methods such as “data bucket-
ing” (NovelAI, 2023) successfully tackle the
same task, we still benefit from cropping-
induced data augmentation, while making sure
that it does not leak into the generation process
- we actually use it to our advantage to gain
more control over the image synthesis process.
Furthermore, it is easy to implement and can
be applied in an online fashion during training,
without additional data preprocessing.
5

Under review as a conference paper at ICLR 2024
ccrop= (0,0) ccrop= (0,256), ccrop= (256,0), ccrop= (512,512),
“An astronaut riding a pig, highly realistic dslr photo, cinematic shot.”“A capybara made of lego sitting in a realistic, natural field.”
Figure 5: Varying the crop conditioning as discussed in Sec. 2.2. See Fig. 4 and Fig. 15 for samples fromSD1.5
andSD2.1 which provide no explicit control of this parameter and thus introduce cropping artifacts. Samples
from the512
2
model, see Sec. 2.5.
2.3 MULTI-ASPECTTRAINING
Real-world datasets include images of widely varying sizes and aspect-ratios (c.f. fig. 2) While the
common output resolutions for text-to-image models are square images of512×512 or1024×1024
pixels, we argue that this is a rather unnatural choice, given the widespread distribution and use of
landscape (e.g., 16:9) or portrait format screens.
Motivated by this observation, we finetune our model to handle multiple aspect-ratios simultaneously:
We follow common practice (NovelAI, 2023) and partition the data into buckets of different aspect
ratios, where we keep the pixel count as close to1024
2 pixels as possibly, varying height and width
accordingly in multiples of 64. A full list of all aspect ratios used for training is provided in App. H.
During optimization, a training batch is composed of images from the same bucket, and we alternate
between bucket sizes for each training step. Additionally, the model receives the bucket size (or,
target size) as a conditioning, represented as a tuple of integerscar= (htgt, wtgt) which are embedded
into a Fourier space similarly to the size- and crop-conditionings described above.
In practice, we apply multi-aspect training as a finetuning stage after pretraining the model at a
fixed aspect-ratio and resolution and combine it with the conditioning techniques introduced in
Sec. 2.2 via concatenation along the channel axis. Fig. 17 in App. I providespython-code for this
operation. Note that crop-conditioning and multi-aspect training are complementary operations, and
crop-conditioning then only works within the bucket boundaries (usually 64 pixels). For ease of
implementation, however, we opt to keep this control parameter for multi-aspect models. We note
that this mechanism can be extended to jointmulti-aspect, multi-resolutiontraining by varying the
total pixel density. In practice, this means that the batch size can be adjusted dynamically depending
on the current resolution, to make the best use of availalable VRAM.
6

Under review as a conference paper at ICLR 2024
2.4 IMPROVEDAUTOENCODER
Table 3: Autoencoder reconstruction performance on
the COCO2017 Lin et al. (2015) validation split, images
of size256×256 pixels. Note:Stable Diffusion2.x uses
an improved version ofStable Diffusion1.x’s autoen-
coder, where the decoder was finetuned with a reduced
weight on the perceptual loss Zhang et al. (2018), and
used more compute. Note that our new autoencoder is
trained from scratch.
model PNSR ↑SSIM↑LPIPS↓rFID↓
SDXL-VAE 24.7 0 .73 0 .88 4 .4
SD-VAE 1.x 23.4 0.69 0.96 5.0
SD-VAE 2.x 24.5 0.71 0.92 4.7
Stable Diffusionis aLDM, operating in a pre-
trained, learned (and fixed) latent space of an
autoencoder (AE). While the bulk of the se-
mantic composition is done by the LDM (Rom-
bach et al., 2021), we can improvelocal, high-
frequency details in generated images by im-
proving the AE. To this end, we train the same
AE architecture used for the originalStable Dif-
fusionat a batch-size of 256 and additionally
track the weights with an exponential moving
average. The resulting AE outperforms the origi-
nal model in all evaluated reconstruction metrics,
see Tab. 3. A small ablation for the influence of
these parameters is reported in App. J, we find
that EMA is helpful in all our settings, while the effects of the large batch size are mixed. We use this
AE for all of our experiments.
2.5 PUTTINGEVERYTHINGTOGETHER
We train the final model,SDXL, in a multi-stage procedure.SDXLuses the autoencoder from Sec. 2.4
and a discrete-time diffusion schedule (Ho et al., 2020; Sohl-Dickstein et al., 2015) with1000steps.
First, we pretrain a base model (see Tab. 1) on an internal dataset whose height- and width-distribution
is visualized in Fig. 2 for600 000optimization steps at a resolution of256×256 pixels and a batch-
size of2048, using size- and crop-conditioning as described in Sec. 2.2. We continue training on
512px for another200 000optimization steps, and finally utilize multi-aspect training (Sec. 2.3) in
combination with an offset-noise (Guttenberg & CrossLabs, 2023; Lin et al., 2023) level of0.05to
train the model on different aspect ratios (Sec. 2.3, App. H) of∼1024×1024pixel area.
Figure 6:1024
2 samples (with zoom-ins) fromSDXLwithout (left) and with (right) therefiner model
(see Sec. 2.5). Prompt:“Epic long distance cityscape photo of New York City flooded by the ocean and
overgrown buildings and jungle ruins in rainforest, at sunset, cinematic shot, highly detailed, 8k, golden light”.
See Fig. 14 for additional samples.
7

Under review as a conference paper at ICLR 2024
Refinement StageEmpirically, we find that the resulting model sometimes yields samples of low
local quality, see Fig. 6. To improve sample quality, we train a separate LDM in the same latent space,
which is specialized on high-quality, high resolution data and employ a noising-denoising process as
introduced bySDEdit(Meng et al., 2021) on the samples from the base model, or, alternatively, finish
the denoising process with the refiner. We follow (Balaji et al., 2022) and specialize this refinement
model on the first 200 (discrete) noise scales. During inference, we render latents from the base
SDXL, and directly diffuse and denoise them in latent space with the refinement model (see Fig. 1),
using the same text input. We note that this step is optional, but improves sample quality for detailed
backgrounds and human faces, as demonstrated in Fig. 6 and Fig. 14.
To assess the performance of our model (with and without refinement stage), we conduct a user
study, and let users pick their favorite generation from the following four models:SDXL,SDXL
(with refiner),Stable Diffusion1.5 andStable Diffusion2.1. The results demonstrate theSDXLwith
the refinement stage is the highest rated choice, and outperformsStable Diffusion1.5 & 2.1 by a
significant margin (win rates:SDXLw/ refinement:48.44%,SDXLbase:36.93%,Stable Diffusion
1.5:7.91%,Stable Diffusion2.1:6.71%). See Fig. 1, which also provides an overview of the full
pipeline. However, when using classical performance metrics such as FID and CLIP scores the
improvements ofSDXLover previous methods are not reflected as shown in Fig. 13 and discussed in
App. E. This aligns with and further backs the findings of Kirstain et al. (2023).
Input Image Input Prompt and Outputs
’Beautiful watercolor painting.’
’An elephant.’
’A house.’ ’Cute animals.’
Figure 7: Adding multimodal control: Replacing the pooled text representations of CLIP Radford et al. (2021),
which were used during training, with CLIP image features turnsSDXLinto a multimodal image generator for
text-controlled image editing, which can even transfer abstract concepts such as "image grid" (bottom row)
from the input image to its output. In contrast to previous work Balaji et al. (2022), this does not require joint
image-text-conditioned pretraining but only 1000 finetuning steps of a single network layer.
2.6 MULTIMODALCONTROL
Starting withSDEdit(Meng et al., 2021), adding image guidance beyond plain text has been a major
focus of numerous recent works (Zhang & Agrawala, 2023; Mou et al., 2023; Ruiz et al., 2023;
Kawar et al., 2023; Hertz et al., 2022), both with and without further training of the base model. In
8

Under review as a conference paper at ICLR 2024
this section, we describe a simple and efficient approach to turningSDXLinto a model guided by
bothtext prompts and input images.
As described in Sec. 2.1, we modify the originalStable Diffusionarchitecture by considering not
only the text embedding sequence, but also the pooled (global) text representation of the CLIP-G
text encoder. By taking advantage of the fact that the pooled CLIP feature space is a (globally)
shared image-text feature space, we can replace this global text representation with a global image
representation from the CLIP-G image encoder. To account for the slight discrepancy between image
and text embeddings, we fine-tune the embedding layer that maps the CLIP embedding to the UNet’s
timestep embedding space (where they are added), and leave the remaining parameters frozen.
Fig. 7 demonstratesSDXL’s multimodal processing capabilities after this fine-tuning, where we
prompt the model with both an input image and text input. For example, the model is able to
extract the concept "grid" from an input image and transfer it to another output controlled by a text
prompt (see bottom row, Fig. 7). We note that a similar approach was implemented in (Balaji et al.,
2022), utilizing joint image and video training (from scratch) with high dropout rates on the image
conditioning. In contrast, using a trained text-to-image model ofSDXL, we can replace the pooled
CLIP embeddings in a zero-shot manner and achieve high quality by finetuning only the embedding
layer for a few thousand steps. Note that we only modify the base model and leave the refiner as is.
3 CONCLUSION& FUTUREWORK
This report presents an analysis of improvements to the foundation modelStable Diffusionfor text-to-
image synthesis. While we achieve significant improvements in synthesized image quality, prompt
adherence and composition, we believe the model may be improved further in the following aspects:
Single stage:Currently, we generate the best samples fromSDXLwith a two-stage approach using
our refinement model. This results in having to load two large models into memory, hampering
accessibility and sampling speed. Future work should investigate ways to provide a single stage.
Text synthesis:While the scale and the larger text encoder (OpenCLIP ViT-bigG (Ilharco et al.,
2021)) help to improve the text rendering capabilities over previous versions ofStable Diffusion,
incorporating byte-level tokenizers (Xue et al., 2022; Liu et al., 2023) or simply scaling the model to
larger sizes (Yu et al., 2022; Saharia et al., 2022) should further improve text synthesis.
Architecture:During the exploration stage of this work, we briefly experimented with transformer-
based architectures such as UViT (Hoogeboom et al., 2023) and DiT (Peebles & Xie, 2022), but
found no immediate benefit. We remain, however, optimistic that a careful hyperparameter study will
eventually enable scaling to much larger transformer-dominated architectures.
Distillation:While our improvements overStable Diffusionare significant, they come at the price
of increased inference cost (both in VRAM and sampling speed). Future work will thus focus on
decreasing the compute needed for inference, and increased sampling speed, for example through
guidance- (Meng et al., 2023), knowledge- (Dockhorn et al., 2023; Kim et al., 2023; Li et al., 2023)
and progressive distillation (Salimans & Ho, 2022; Berthelot et al., 2023; Meng et al., 2023).
Finally, our model is trained in the discrete-time formulation of (Ho et al., 2020), and requires
offset-noise(Guttenberg & CrossLabs, 2023; Lin et al., 2023) for aesthetically pleasing results.
The EDM-framework of Karras et al. (2022) is a promising candidate for future model training, as
its formulation in continuous time allows for increased sampling flexibility and does not require
noise-schedule corrections.
9

Under review as a conference paper at ICLR 2024
REFERENCES
Yogesh Balaji, Seungjun Nah, Xun Huang, Arash Vahdat, Jiaming Song, Karsten Kreis, Miika Aittala,
Timo Aila, Samuli Laine, Bryan Catanzaro, Tero Karras, and Ming-Yu Liu. eDiff-I: Text-to-Image
Diffusion Models with an Ensemble of Expert Denoisers.arXiv:2211.01324, 2022.
David Berthelot, Arnaud Autef, Jierui Lin, Dian Ang Yap, Shuangfei Zhai, Siyuan Hu, Daniel
Zheng, Walter Talbot, and Eric Gu. TRACT: Denoising Diffusion Models with Transitive Closure
Time-Distillation.arXiv:2303.04248, 2023.
Andreas Blattmann, Robin Rombach, Huan Ling, Tim Dockhorn, Seung Wook Kim, Sanja Fidler,
and Karsten Kreis. Align your Latents: High-Resolution Video Synthesis with Latent Diffusion
Models.arXiv:2304.08818, 2023.
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale
hierarchical image database. In2009 IEEE conference on computer vision and pattern recognition,
pp. 248–255. Ieee, 2009.
Prafulla Dhariwal and Alex Nichol. Diffusion Models Beat GANs on Image Synthesis.
arXiv:2105.05233, 2021.
Tim Dockhorn, Robin Rombach, Andreas Blattmann, and Yaoliang Yu. Distilling the Knowledge in
Diffusion Models.CVPR Workshop on Generative Models for Computer Vision, 2023.
Patrick Esser, Johnathan Chiu, Parmida Atighehchian, Jonathan Granskog, and Anastasis Germanidis.
Structure and content-guided video synthesis with diffusion models, 2023.
Weixi Feng, Xuehai He, Tsu-Jui Fu, Varun Jampani, Arjun Akula, Pradyumna Narayana, Sugato
Basu, Xin Eric Wang, and William Yang Wang. Training-free structured diffusion guidance for
compositional text-to-image synthesis.arXiv:2212.05032, 2023.
Seth Forsgren and Hayk Martiros. Riffusion - Stable diffusion for real-time music generation, 2022.
URLhttps://riffusion.com/about .
Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H Bermano, Gal Chechik, and Daniel
Cohen-Or. An image is worth one word: Personalizing text-to-image generation using textual
inversion.arXiv:2208.01618, 2022.
Nicholas Guttenberg and CrossLabs. Diffusion with offset noise, 2023. URLhttps://www.
crosslabs.org/blog/diffusion-with-offset-noise .
Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Prompt-
to-prompt image editing with cross attention control, 2022.
Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter.
GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium.
arXiv:1706.08500, 2017.
Jonathan Ho and Tim Salimans. Classifier-Free Diffusion Guidance.arXiv:2207.12598, 2022.
Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising Diffusion Probabilistic Models.arXiv preprint
arXiv:2006.11239, 2020.
Jonathan Ho, William Chan, Chitwan Saharia, Jay Whang, Ruiqi Gao, Alexey Gritsenko, Diederik P
Kingma, Ben Poole, Mohammad Norouzi, David J Fleet, and Tim Salimans. Imagen Video: High
Definition Video Generation with Diffusion Models.arXiv:2210.02303, 2022.
David Holz. Midjourney, 2023. URLhttps://www.midjourney.com/home/ .
Emiel Hoogeboom, Jonathan Heek, and Tim Salimans. simple diffusion: End-to-end diffusion for
high resolution images.arXiv preprint arXiv:2301.11093, 2023.
Rongjie Huang, Jiawei Huang, Dongchao Yang, Yi Ren, Luping Liu, Mingze Li, Zhenhui Ye, Jinglin
Liu, Xiang Yin, and Zhou Zhao. Make-An-Audio: Text-To-Audio Generation with Prompt-
Enhanced Diffusion Models.arXiv:2301.12661, 2023.
10

Under review as a conference paper at ICLR 2024
Aapo Hyvärinen and Peter Dayan. Estimation of Non-Normalized Statistical Models by Score
Matching.Journal of Machine Learning Research, 6(4), 2005.
Gabriel Ilharco, Mitchell Wortsman, Ross Wightman, Cade Gordon, Nicholas Carlini, Rohan Taori,
Achal Dave, Vaishaal Shankar, Hongseok Namkoong, John Miller, Hannaneh Hajishirzi, Ali
Farhadi, and Ludwig Schmidt. OpenCLIP, July 2021. URLhttps://doi.org/10.5281/
zenodo.5143773.
Heewoo Jun, Rewon Child, Mark Chen, John Schulman, Aditya Ramesh, Alec Radford, and Ilya
Sutskever. Distribution Augmentation for Generative Modeling. InInternational Conference on
Machine Learning, pp. 5006–5019. PMLR, 2020.
Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the Design Space of Diffusion-
Based Generative Models.arXiv:2206.00364, 2022.
Bahjat Kawar, Shiran Zada, Oran Lang, Omer Tov, Huiwen Chang, Tali Dekel, Inbar Mosseri, and
Michal Irani. Imagic: Text-based real image editing with diffusion models, 2023.
Bo-Kyeong Kim, Hyoung-Kyu Song, Thibault Castells, and Shinkook Choi. On Architectural
Compression of Text-to-Image Diffusion Models.arXiv:2305.15798, 2023.
Yuval Kirstain, Adam Polyak, Uriel Singer, Shahbuland Matiana, Joe Penna, and Omer Levy. Pick-
a-pic: An open dataset of user preferences for text-to-image generation.arXiv:2305.01569,
2023.
Yanyu Li, Huan Wang, Qing Jin, Ju Hu, Pavlo Chemerys, Yun Fu, Yanzhi Wang, Sergey Tulyakov,
and Jian Ren. SnapFusion: Text-to-Image Diffusion Model on Mobile Devices within Two Seconds.
arXiv:2306.00980, 2023.
Shanchuan Lin, Bingchen Liu, Jiashi Li, and Xiao Yang. Common Diffusion Noise Schedules and
Sample Steps are Flawed.arXiv:2305.08891, 2023.
Tsung-Yi Lin, Michael Maire, Serge Belongie, Lubomir Bourdev, Ross Girshick, James Hays, Pietro
Perona, Deva Ramanan, C. Lawrence Zitnick, and Piotr Dollár. Microsoft coco: Common objects
in context, 2015.
Rosanne Liu, Dan Garrette, Chitwan Saharia, William Chan, Adam Roberts, Sharan Narang, Irina
Blok, RJ Mical, Mohammad Norouzi, and Noah Constant. Character-aware models improve visual
text rendering, 2023.
Chenlin Meng, Yang Song, Jiaming Song, Jiajun Wu, Jun-Yan Zhu, and Stefano Ermon. SDEdit:
Guided Image Synthesis and Editing with Stochastic Differential Equations.arXiv:2108.01073,
2021.
Chenlin Meng, Robin Rombach, Ruiqi Gao, Diederik P. Kingma, Stefano Ermon, Jonathan Ho, and
Tim Salimans. On distillation of guided diffusion models, 2023.
Chong Mou, Xintao Wang, Liangbin Xie, Yanze Wu, Jian Zhang, Zhongang Qi, Ying Shan, and
Xiaohu Qie. T2i-adapter: Learning adapters to dig out more controllable ability for text-to-image
diffusion models, 2023.
Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya
Sutskever, and Mark Chen. GLIDE: Towards Photorealistic Image Generation and Editing with
Text-Guided Diffusion Models.arXiv:2112.10741, 2021.
NovelAI. Novelai improvements on stable diffusion, 2023. URLhttps://blog.novelai.
net/novelai-improvements-on-stable-diffusion-e10d38db82ac .
Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor
Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Köpf, Edward
Yang, Zach DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang,
Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning
library, 2019.
11

Under review as a conference paper at ICLR 2024
William Peebles and Saining Xie. Scalable Diffusion Models with Transformers.arXiv:2212.09748,
2022.
Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal,
Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever.
Learning Transferable Visual Models From Natural Language Supervision.arXiv:2103.00020,
2021.
Aditya Ramesh. How dall·e 2 works, 2022. URLhttp://adityaramesh.com/posts/
dalle2/dalle2.html.
Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen,
and Ilya Sutskever. Zero-shot text-to-image generation, 2021.
Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical Text-
Conditional Image Generation with CLIP Latents.arXiv:2204.06125, 2022.
Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-
Resolution Image Synthesis with Latent Diffusion Models.arXiv preprint arXiv:2112.10752,
2021.
Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-Net: Convolutional Networks for Biomedical
Image Segmentation.arXiv:1505.04597, 2015.
Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman.
Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation, 2023.
Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed Kamyar
Seyed Ghasemipour, Burcu Karagol Ayan, S Sara Mahdavi, Rapha Gontijo Lopes, Tim Salimans,
Jonathan Ho, David J Fleet, and Mohammad Norouzi. Photorealistic Text-to-Image Diffusion
Models with Deep Language Understanding.arXiv:2205.11487, 2022.
Tim Salimans and Jonathan Ho. Progressive Distillation for Fast Sampling of Diffusion Models.
arXiv preprint arXiv:2202.00512, 2022.
Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen.
Improved Techniques for Training GANs.arXiv:1606.03498, 2016.
Sitian Shen, Zilin Zhu, Linqian Fan, Harry Zhang, and Xinxiao Wu. DiffCLIP: Leveraging Stable
Diffusion for Language Grounded 3D Classification.arXiv:2305.15957, 2023.
Uriel Singer, Adam Polyak, Thomas Hayes, Xi Yin, Jie An, Songyang Zhang, Qiyuan Hu, Harry
Yang, Oron Ashual, Oran Gafni, Devi Parikh, Sonal Gupta, and Yaniv Taigman. Make-A-Video:
Text-to-Video Generation without Text-Video Data.arXiv:2209.14792, 2022.
Jascha Sohl-Dickstein, Eric A Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep Unsupervised
Learning using Nonequilibrium Thermodynamics.arXiv:1503.03585, 2015.
Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models.
arXiv:2010.02502, 2020a.
Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and
Ben Poole. Score-Based Generative Modeling through Stochastic Differential Equations.
arXiv:2011.13456, 2020b.
Andreas Stöckl. Evaluating a synthetic image dataset generated with stable diffusion.
arXiv:2211.01777, 2022.
Yu Takagi and Shinji Nishimoto. High-Resolution Image Reconstruction With Latent Diffusion
Models From Human Brain Activity. InProceedings of the IEEE/CVF Conference on Computer
Vision and Pattern Recognition, pp. 14453–14463, 2023.
12

Under review as a conference paper at ICLR 2024
Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée
Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand
Joulin, Edouard Grave, and Guillaume Lample. LLaMA: Open and Efficient Foundation Language
Models.arXiv:2302.13971, 2023.
Jialiang Wei, Anne-Lise Courbis, Thomas Lambolais, Binbin Xu, Pierre Louis Bernard, and Gérard
Dray. Boosting gui prototyping with diffusion models.arXiv preprint arXiv:2306.06233, 2023.
Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam
Roberts, and Colin Raffel. Byt5: Towards a token-free future with pre-trained byte-to-byte models,
2022.
Jiahui Yu, Yuanzhong Xu, Jing Yu Koh, Thang Luong, Gunjan Baid, Zirui Wang, Vijay Vasudevan,
Alexander Ku, Yinfei Yang, Burcu Karagol Ayan, Ben Hutchinson, Wei Han, Zarana Parekh, Xin
Li, Han Zhang, Jason Baldridge, and Yonghui Wu. Scaling autoregressive models for content-rich
text-to-image generation, 2022.
Lvmin Zhang and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models.
arXiv:2302.05543, 2023.
Richard Zhang, Phillip Isola, Alexei A. Efros, Eli Shechtman, and Oliver Wang. The unreasonable
effectiveness of deep features as a perceptual metric, 2018.
13