Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
140 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Tamed Unadjusted Langevin Algorithm (1710.05559v3)

Published 16 Oct 2017 in stat.ME

Abstract: In this article, we consider the problem of sampling from a probability measure $\pi$ having a density on $\mathbb{R}d$ known up to a normalizing constant, $x\mapsto \mathrm{e}{-U(x)} / \int_{\mathbb{R}d} \mathrm{e}{-U(y)} \mathrm{d} y$. The Euler discretization of the Langevin stochastic differential equation (SDE) is known to be unstable in a precise sense, when the potential $U$ is superlinear, i.e. $\liminf_{\Vert x \Vert\to+\infty} \Vert \nabla U(x) \Vert / \Vert x \Vert = +\infty$. Based on previous works on the taming of superlinear drift coefficients for SDEs, we introduce the Tamed Unadjusted Langevin Algorithm (TULA) and obtain non-asymptotic bounds in $V$-total variation norm and Wasserstein distance of order $2$ between the iterates of TULA and $\pi$, as well as weak error bounds. Numerical experiments are presented which support our findings.

Summary

We haven't generated a summary for this paper yet.