Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled Markov Chains (2010.01845v2)

Published 5 Oct 2020 in cs.LG and stat.ML

Abstract: The variational auto-encoder (VAE) is a deep latent variable model that has two neural networks in an autoencoder-like architecture; one of them parameterizes the model's likelihood. Fitting its parameters via maximum likelihood (ML) is challenging since the computation of the marginal likelihood involves an intractable integral over the latent space; thus the VAE is trained instead by maximizing a variational lower bound. Here, we develop a ML training scheme for VAEs by introducing unbiased estimators of the log-likelihood gradient. We obtain the estimators by augmenting the latent space with a set of importance samples, similarly to the importance weighted auto-encoder (IWAE), and then constructing a Markov chain Monte Carlo coupling procedure on this augmented space. We provide the conditions under which the estimators can be computed in finite time and with finite variance. We show experimentally that VAEs fitted with unbiased estimators exhibit better predictive performance.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Francisco J. R. Ruiz (22 papers)
  2. Michalis K. Titsias (39 papers)
  3. Taylan Cemgil (10 papers)
  4. Arnaud Doucet (161 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.