Critical windows: non-asymptotic theory for feature emergence in diffusion models (2403.01633v2)
Abstract: We develop theory to understand an intriguing property of diffusion models for image generation that we term critical windows. Empirically, it has been observed that there are narrow time intervals in sampling during which particular features of the final image emerge, e.g. the image class or background color (Ho et al., 2020b; Meng et al., 2022; Choi et al., 2022; Raya & Ambrogioni, 2023; Georgiev et al., 2023; Sclocchi et al., 2024; Biroli et al., 2024). While this is advantageous for interpretability as it implies one can localize properties of the generation to a small segment of the trajectory, it seems at odds with the continuous nature of the diffusion. We propose a formal framework for studying these windows and show that for data coming from a mixture of strongly log-concave densities, these windows can be provably bounded in terms of certain measures of inter- and intra-group separation. We also instantiate these bounds for concrete examples like well-conditioned Gaussian mixtures. Finally, we use our bounds to give a rigorous interpretation of diffusion models as hierarchical samplers that progressively "decide" output features over a discrete sequence of times. We validate our bounds with synthetic experiments. Additionally, preliminary experiments on Stable Diffusion suggest critical windows may serve as a useful tool for diagnosing fairness and privacy violations in real-world diffusion models.
- Linear convergence bounds for diffusion models via stochastic localization. arXiv preprint arXiv:2308.03686, 2023a.
- Error bounds for flow matching methods. arXiv preprint arXiv:2305.16860, 2023b.
- Dynamical regimes of diffusion models, 2024.
- Generative modeling with denoising auto-encoders and Langevin sampling. arXiv preprint 2002.00107, 2022.
- Extracting training data from diffusion models. In Proceedings of the 32nd USENIX Conference on Security Symposium, SEC ’23, USA, 2023. USENIX Association. ISBN 978-1-939133-37-3.
- Improved analysis of score-based generative modeling: user-friendly bounds under minimal smoothness assumptions. arXiv preprint arXiv:2211.01916, 2022.
- Improved analysis of score-based generative modeling: User-friendly bounds under minimal smoothness assumptions. In International Conference on Machine Learning, pp. 4735–4763. PMLR, 2023a.
- The probability flow ode is provably fast. arXiv preprint arXiv:2305.11798, 2023b.
- Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net, 2023c. URL https://openreview.net/pdf?id=zyLVMgsZ0U_.
- Restoration-degradation beyond linear diffusions: A non-asymptotic analysis for ddim-type samplers. arXiv preprint arXiv:2303.03384, 2023d.
- Analysis of learning a flow-based generative model from limited sample complexity. arXiv preprint arXiv:2310.03575, 2023.
- De Bortoli, V. Convergence of denoising diffusion models under the manifold hypothesis. Transactions on Machine Learning Research, 2022.
- Diffusion Schrödinger bridge with applications to score-based generative modeling. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 17695–17709. Curran Associates, Inc., 2021.
- Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780–8794, 2021.
- Are diffusion models vulnerable to membership inference attacks? In Proceedings of the 40th International Conference on Machine Learning, ICML’23. JMLR.org, 2023.
- Interpreting clip’s image representation via text-based decomposition, 2024.
- The journey, not the destination: How data guides diffusion models. arXiv preprint arXiv:2312.06205, 2023.
- Log-concave observers. In 17th International Symposium on Mathematical Theory of Networks and Systems, 2006, 2006. 17th International Symposium on Mathematical Theory of Networks and Systems, 2006 : MTNS 2006 ; Conference date: 24-07-2006 Through 28-07-2006.
- Denoising diffusion probabilistic models. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 6840–6851. Curran Associates, Inc., 2020a. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/4c5bcfec8584af0d967f1ab10179ca4b-Paper.pdf.
- Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851, 2020b. URL https://arxiv.org/abs/2006.11239.
- Isoperimetric problems for convex bodies and a localization lemma. Discrete & Computational Geometry, 13:541–559, 1995.
- Sampling multimodal distributions with the vanilla score: Benefits of data-based initialization. arXiv preprint arXiv:2310.01762, 2023.
- LeCam, L. Asymptotic methods in statistical decision theory. Springer series in statistics. Springer, New York, NY [u.a.], 1986. ISBN 3540963073. URL http://gso.gbv.de/DB=2.1/CMD?ACT=SRCHA&SRT=YOP&IKT=1016&TRM=ppn+024181773&sourceid=fbw_bibsonomy.
- Beyond log-concavity: Provable guarantees for sampling multi-modal distributions using simulated tempering langevin monte carlo. Advances in neural information processing systems, 31, 2018.
- Convergence for score-based generative modeling with polynomial complexity. In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K. (eds.), Advances in Neural Information Processing Systems, 2022.
- Convergence of score-based generative modeling for general data distributions. In International Conference on Algorithmic Learning Theory, pp. 946–985. PMLR, 2023.
- Towards faster non-asymptotic convergence for diffusion-based generative models. arXiv preprint arXiv:2306.09251, 2023a.
- Towards a mathematical theory for consistency training in diffusion models. arXiv preprint arXiv:2402.07802, 2024.
- Zero-shot machine-generated image detection using sinks of gradient flows. https://github.com/deep-learning-mit/staging/blob/main/_posts/2023-11-08-detect-image.md, 2023.
- MoPe: Model perturbation based privacy attacks on language models. In Bouamor, H., Pino, J., and Bali, K. (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 13647–13660, Singapore, December 2023b. Association for Computational Linguistics. doi: 10.18653/v1/2023.emnlp-main.842. URL https://aclanthology.org/2023.emnlp-main.842.
- Li, S. Concise formulas for the area and volume of a hyperspherical cap. Asian J. Math. Stat., 4(1):66–70, December 2010.
- Let us build bridges: understanding and extending diffusion generative models. arXiv preprint arXiv:2208.14699, 2022.
- Stable bias: Analyzing societal representations in diffusion models, 2023.
- Detectgpt: zero-shot machine-generated text detection using probability curvature. In Proceedings of the 40th International Conference on Machine Learning, ICML’23. JMLR.org, 2023.
- Pardo, L. Statistical Inference Based on Divergence Measures. CRC Press, Abingdon, 2005. URL https://cds.cern.ch/record/996837.
- How deep neural networks learn compositional data: The random hierarchy model. arXiv preprint arXiv:2307.02129, 2023.
- Pidstrigach, J. Score-based generative models detect manifolds. In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (eds.), Advances in Neural Information Processing Systems, volume 35, pp. 35852–35865. Curran Associates, Inc., 2022.
- Learning transferable visual models from natural language supervision, 2021.
- Spontaneous symmetry breaking in generative diffusion models. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=lxGFGMMSVl.
- High-dimensional statistics, 2023.
- Log-concavity and strong log-concavity: A review. Statistics Surveys, 8(none):45 – 114, 2014. doi: 10.1214/14-SS107. URL https://doi.org/10.1214/14-SS107.
- A phase transition in diffusion models reveals the hierarchical nature of data, 2024.
- Learning mixtures of gaussians using the ddpm objective. arXiv preprint arXiv:2307.01178, 2023.
- Membership inference attacks against machine learning models. In 2017 IEEE Symposium on Security and Privacy, SP 2017, San Jose, CA, USA, May 22-26, 2017, pp. 3–18. IEEE Computer Society, 2017. doi: 10.1109/SP.2017.41. URL https://doi.org/10.1109/SP.2017.41.
- Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp. 2256–2265. PMLR, 2015.
- Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456, 2020.
- Convergence in KL divergence of the inexact Langevin algorithm with application to score-based generative models. arXiv preprint 2211.01512, 2022.
- Marvin Li (5 papers)
- Sitan Chen (57 papers)