Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fast Sampling via Discrete Non-Markov Diffusion Models with Predetermined Transition Time (2312.09193v3)

Published 14 Dec 2023 in cs.LG, cs.AI, and stat.ML

Abstract: Discrete diffusion models have emerged as powerful tools for high-quality data generation. Despite their success in discrete spaces, such as text generation tasks, the acceleration of discrete diffusion models remains under-explored. In this paper, we propose discrete non-Markov diffusion models (DNDM), which naturally induce the predetermined transition time set. This enables a training-free sampling algorithm that significantly reduces the number of function evaluations (i.e., calls to the neural network), making the sampling process much faster. Furthermore, we study the transition from finite to infinite step sampling, offering new insights into bridging the gap between discrete and continuous-time processes for discrete diffusion models. Extensive experiments on natural language generation and machine translation tasks demonstrate the superior performance of our method in terms of both generation speed and sample quality compared to existing methods for discrete diffusion models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zixiang Chen (28 papers)
  2. Huizhuo Yuan (16 papers)
  3. Yongqian Li (6 papers)
  4. Yiwen Kou (6 papers)
  5. Junkai Zhang (25 papers)
  6. Quanquan Gu (198 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.