Generalization of Diffusion Models Arises from a Regularized Representation Space

Published: 23 Sept 2025, Last Modified: 23 Dec 2025SPIGM @ NeurIPSEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Diffusion models; Memorization; Generalization; Representation learning; Autoencoders
TL;DR: Learning good representations is central to novel and meaningful image generation with diffusion models.
Abstract: Diffusion models generate high-quality, diverse samples with great generalizability, yet when overfit to the training objective, they may memorize training data. We analyze memorization and generalization of diffusion models through the lens of representation learning. Using a two-layer ReLU denoising autoencoder (DAE) parameterization, we prove that memorization corresponds to the model learning the raw data matrix for encoding and decoding, yielding spiky representations; in contrast, generalization arises when the model captures local data statistics, producing balanced representations. We validate these insights by investigating representation spaces in real-world unconditional and text-to-image diffusion models, where the same distinctions emerge. Practically, we propose a representation-based memorization detection method and a training-free editing method that allows precise control via representation steering. Together, our results underscore that learning good representations is central to novel and meaningful generation.
Submission Number: 75
Loading