Interpretable Diffusion via Information Decomposition (2310.07972v3)
Abstract: Denoising diffusion models enable conditional generation and density modeling of complex relationships like images and text. However, the nature of the learned relationships is opaque making it difficult to understand precisely what relationships between words and parts of an image are captured, or to predict the effect of an intervention. We illuminate the fine-grained relationships learned by diffusion models by noticing a precise relationship between diffusion and information decomposition. Exact expressions for mutual information and conditional mutual information can be written in terms of the denoising model. Furthermore, pointwise estimates can be easily estimated as well, allowing us to ask questions about the relationships between specific images and captions. Decomposing information even further to understand which variables in a high-dimensional space carry information is a long-standing problem. For diffusion models, we show that a natural non-negative decomposition of mutual information emerges, allowing us to quantify informative relationships between words and pixels in an image. We exploit these new relations to measure the compositional understanding of diffusion models, to do unsupervised localization of objects in images, and to measure effects when selectively editing images through prompt interventions.
- ediffi: Text-to-image diffusion models with an ensemble of expert denoisers. arXiv preprint arXiv:2211.01324, 2022.
- Clustering with bregman divergences. Journal of machine learning research, 6(10), 2005.
- Identifying and mitigating the security risks of generative ai. arXiv preprint arXiv:2308.14840, 2023.
- Mutual information neural estimation. In Proceedings of the 35th International Conference on Machine Learning, pp. 531–540, 2018.
- Easily accessible text-to-image generation amplifies demographic stereotypes at large scale. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pp. 1493–1504, 2023.
- Multimodal datasets: misogyny, pornography, and malignant stereotypes. arXiv preprint arXiv:2110.01963, 2021.
- Improving mutual information estimation with annealed and energy-based bounds. arXiv preprint arXiv:2303.06992, 2023.
- Diffusion policy: Visuomotor policy learning via action diffusion. arXiv preprint arXiv:2303.04137, 2023.
- Testing relational understanding in text-guided image generation. arXiv preprint arXiv:2208.00005, 2022.
- Elements of information theory. Wiley-Interscience, 2006.
- Robert M Fano. Transmission of Information: A Statistical Theory of Communications. MIT Press, 1961.
- Pointwise partial information decompositionusing the specificity and ambiguity lattices. Entropy, 20(4):297, 2018.
- Mutual information and minimum mean-square error in gaussian channels. IEEE transactions on information theory, 51(4):1261–1282, 2005.
- How does gpt-2 compute greater-than?: Interpreting mathematical abilities in a pre-trained language model. arXiv preprint arXiv:2305.00586, 2023.
- Discriminative diffusion models as few-shot vision and language learners, 2023.
- An overview of catastrophic ai risks. arXiv preprint arXiv:2306.12001, 2023.
- Equivariant 3d-conditional diffusion models for molecular linker design. arXiv preprint arXiv:2210.05274, 2022.
- Openclip, July 2021. URL https://doi.org/10.5281/zenodo.5143773.
- Ddp: Diffusion model for dense visual prediction, 2023.
- Diffusion models for zero-shot open-vocabulary segmentation. arXiv preprint arXiv:2306.09316, 2023.
- Elucidating the design space of diffusion-based generative models. Advances in Neural Information Processing Systems, 35:26565–26577, 2022.
- Imagic: Text-based real image editing with diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6007–6017, 2023.
- Steven M Kay. Fundamentals of statistical signal processing: estimation theory. Prentice-Hall, Inc., 1993.
- Variational diffusion models. arXiv preprint arXiv:2107.00630, 2021.
- Redundant information neural estimation. Entropy, 23(7):922, 2021.
- Artemy Kolchinsky. A novel approach to the partial information decomposition. Entropy, 24(3):403, 2022.
- Information-theoretic diffusion. In The Eleventh International Conference on Learning Representations (ICLR), 2022.
- Are diffusion models vision-and-language reasoners? In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
- Maskdiff: Modeling mask distribution with diffusion probabilistic model for few-shot instance segmentation, 2023.
- Neural word embedding as implicit matrix factorization. In Advances in Neural Information Processing Systems, pp. 2177–2185, 2014.
- Factorized contrastive learning: Going beyond multi-view redundancy. arXiv preprint arXiv:2306.05268, 2023.
- Microsoft coco: Common objects in context, 2015.
- Feature pyramid networks for object detection, 2017.
- Zachary C Lipton. The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue, 16(3):31–57, 2018.
- Compositional visual generation with composable diffusion models. In European Conference on Computer Vision, pp. 423–439. Springer, 2022.
- Unsupervised compositional concepts discovery with text-to-image generative models. arXiv preprint arXiv:2306.05357, 2023.
- A unified approach to interpreting model predictions, 2017.
- Diffusionseg: Adapting diffusion towards unsupervised object discovery, 2023.
- Guided image synthesis via initial image editing in diffusion model. arXiv preprint arXiv:2305.03382, 2023.
- David McAllester. On the mathematics of diffusion models. arXiv preprint arXiv:2301.11108, 2023.
- Null-text inversion for editing real images using guided diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6038–6047, 2023.
- Glide: Towards photorealistic image generation and editing with text-guided diffusion models. arXiv preprint arXiv:2112.10741, 2021.
- Mitigating dataset harms requires stewardship: Lessons from 1000 papers. arXiv preprint arXiv:2108.02922, 2021.
- Comprehensive discovery of subsample gene expression components by information explanation: therapeutic implications in cancer. BMC medical genomics, 10(1):12, 2017. URL https://doi.org/10.1186/s12920-017-0245-6.
- Billy Perrigo. Openai used kenyan workers on less than $2 per hour: Exclusive, Jan 2023.
- On variational bounds of mutual information. In Proceedings of the 36th International Conference on Machine Learning, 2019.
- Learning transferable visual models from natural language supervision, 2021.
- Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125, 2022.
- Linguistic binding in diffusion models: Enhancing attribute correspondence through attention map alignment. arXiv preprint arXiv:2306.08877, 2023.
- Influence decompositions for neural network attribution. In The 24th International Conference on Artificial Intelligence and Statistics (AISTATS), 2021.
- High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684–10695, 2022.
- Photorealistic text-to-image diffusion models with deep language understanding. arXiv preprint arXiv:2205.11487, 2022.
- Laion-400m: Open dataset of clip-filtered 400 million image-text pairs. arXiv preprint arXiv:2111.02114, 2021.
- C.E. Shannon. A mathematical theory of communication. The Bell System Technical Journal, 27:379–423, 1948.
- Deep unsupervised learning using nonequilibrium thermodynamics. arXiv preprint arXiv:1503.03585, 2015.
- Denoising diffusion implicit models, 2022.
- Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456, 2020.
- Dual diffusion implicit bridges for image-to-image translation. arXiv preprint arXiv:2203.08382, 2022.
- Axiomatic attribution for deep networks, 2017.
- Intriguing properties of neural networks. In ICLR, 2014.
- Diffss: Diffusion model for few-shot semantic segmentation, 2023.
- What the daam: Interpreting stable diffusion using cross attention, 2022.
- Diffuse, attend, and segment: Unsupervised zero-shot segmentation using stable diffusion, 2023.
- Diffusion model is secretly a training-free open vocabulary semantic segmenter, 2023.
- Interpretability in the wild: a circuit for indirect object identification in gpt-2 small. arXiv preprint arXiv:2211.00593, 2022.
- De novo design of protein structure and function with rfdiffusion. Nature, pp. 1–3, 2023.
- Nonnegative decomposition of multivariate information. arXiv:1004.2515, 2010.
- Open-vocabulary panoptic segmentation with text-to-image diffusion models, 2023.
- When and why vision-language models behave like bag-of-words models, and what to do about it? arXiv preprint arXiv:2210.01936, 2022.
- Diffusionengine: Diffusion model is scalable data engine for object detection, 2023.
- Unleashing text-to-image diffusion models for visual perception, 2023.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Collections
Sign up for free to add this paper to one or more collections.