Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey of Diffusion Models in Natural Language Processing (2305.14671v2)

Published 24 May 2023 in cs.CL

Abstract: This survey paper provides a comprehensive review of the use of diffusion models in NLP. Diffusion models are a class of mathematical models that aim to capture the diffusion of information or signals across a network or manifold. In NLP, diffusion models have been used in a variety of applications, such as natural language generation, sentiment analysis, topic modeling, and machine translation. This paper discusses the different formulations of diffusion models used in NLP, their strengths and limitations, and their applications. We also perform a thorough comparison between diffusion models and alternative generative models, specifically highlighting the autoregressive (AR) models, while also examining how diverse architectures incorporate the Transformer in conjunction with diffusion models. Compared to AR models, diffusion models have significant advantages for parallel generation, text interpolation, token-level controls such as syntactic structures and semantic contents, and robustness. Exploring further permutations of integrating Transformers into diffusion models would be a valuable pursuit. Also, the development of multimodal diffusion models and large-scale diffusion LLMs with notable capabilities for few-shot learning would be important directions for the future advance of diffusion models in NLP.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. Structured denoising diffusion models in discrete state-spaces.
  2. Structured denoising diffusion models in discrete state-spaces. In Advances in Neural Information Processing Systems.
  3. Structured denoising diffusion models in discrete state-spaces. ArXiv, abs/2107.03006.
  4. A neural probabilistic language model. In Journal of machine learning research.
  5. Language models are few-shot learners.
  6. A cheaper and better diffusion language model with soft-masked noise.
  7. Analog bits: Generating discrete data using diffusion models with self-conditioning. In International Conference on Learning Representations.
  8. Bert: Pre-training of deep bidirectional transformers for language understanding.
  9. Prafulla Dhariwal and Alex Nichol. 2021. Diffusion models beat gans on image synthesis.
  10. Continuous diffusion for categorical data.
  11. Difformer: Empowering diffusion models on the embedding space for text generation.
  12. Diffuseq: Sequence to sequence text generation with diffusion models.
  13. Generative adversarial networks.
  14. Ssd-lm: Semi-autoregressive simplex-based diffusion language model for text generation and modular control.
  15. Diffusionbert: Improving generative masked language models with diffusion models.
  16. Denoising diffusion probabilistic models.
  17. Training compute-optimal large language models. ArXiv, abs/2203.15556.
  18. Argmax flows and multinomial diffusion: Learning categorical distributions.
  19. On density estimation with diffusion models. In Advances in Neural Information Processing Systems.
  20. Diederik P Kingma and Max Welling. 2022. Auto-encoding variational bayes.
  21. Durk P Kingma and Prafulla Dhariwal. 2018. Glow: Generative flow with invertible 1x1 convolutions. Advances in neural information processing systems, 31.
  22. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension.
  23. Diffusion-lm improves controllable text generation.
  24. Text generation with diffusion language models: A pre-training approach with continuous paragraph denoise.
  25. Latent diffusion for language generation.
  26. Eliya Nachmani and Shaked Dovrat. 2021. Zero-shot translation using diffusion models.
  27. A systematic characterization of sampling algorithms for open-ended language generation. In Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, pages 334–346, Suzhou, China. Association for Computational Linguistics.
  28. Alex Nichol and Prafulla Dhariwal. 2021. Improved denoising diffusion probabilistic models.
  29. Masked autoregressive flow for density estimation. Advances in neural information processing systems, 30.
  30. Mauve: Measuring the gap between neural text and human text using divergence frontiers.
  31. Pre-trained models for natural language processing: A survey. Science China Technological Sciences, 63(10):1872–1897.
  32. Exploring the limits of transfer learning with a unified text-to-text transformer.
  33. DiffusER: Diffusion via edit-based reconstruction. In International Conference on Learning Representations.
  34. Stochastic backpropagation and approximate inference in deep generative models. In Proceedings of the 31st International Conference on Machine Learning, volume 32 of Proceedings of Machine Learning Research, pages 1278–1286, Bejing, China. PMLR.
  35. High-resolution image synthesis with latent diffusion models.
  36. Step-unrolled denoising autoencoders for text generation.
  37. Deep unsupervised learning using nonequilibrium thermodynamics. In Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pages 2256–2265, Lille, France. PMLR.
  38. Deep unsupervised learning using nonequilibrium thermodynamics.
  39. Denoising diffusion implicit models.
  40. Self-conditioned embedding diffusion for text generation.
  41. Generating text with recurrent neural networks. In International Conference on Machine Learning.
  42. Dinoiser: Diffused conditional sequence learning by manipulating noises.
  43. Seqdiffuseq: Text diffusion with encoder-decoder transformers.
  44. Diffusum: Generation enhanced extractive summarization with diffusion.
  45. Trading off diversity and quality in natural language generation. In Proceedings of the Workshop on Human Evaluation of NLP Systems (HumEval), pages 25–33, Online. Association for Computational Linguistics.
  46. A reparameterized discrete diffusion model for text generation.
Citations (11)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com