Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Denoising Criterion for Variational Auto-Encoding Framework (1511.06406v2)

Published 19 Nov 2015 in cs.LG

Abstract: Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at the input level, while variational autoencoders (VAE) are trained with noise injected in their stochastic hidden layer, with a regularizer that encourages this noise injection. In this paper, we show that injecting noise both in input and in the stochastic hidden layer can be advantageous and we propose a modified variational lower bound as an improved objective function in this setup. When input is corrupted, then the standard VAE lower bound involves marginalizing the encoder conditional distribution over the input noise, which makes the training criterion intractable. Instead, we propose a modified training criterion which corresponds to a tractable bound when input is corrupted. Experimentally, we find that the proposed denoising variational autoencoder (DVAE) yields better average log-likelihood than the VAE and the importance weighted autoencoder on the MNIST and Frey Face datasets.

Citations (189)

Summary

  • The paper presents the denoising variational lower bound, a novel objective that enables efficient VAE training with input noise.
  • It overcomes intractability issues in traditional VAEs by integrating noise at both input and latent layers, yielding improved log-likelihood scores.
  • Empirical results on MNIST and Frey Face datasets demonstrate the robustness and superior performance of the proposed DVAE framework.

An Analysis of the Denoising Criterion for the Variational Auto-Encoding Framework

The paper presents an investigation into a denoising criterion within the context of a variational auto-encoding framework, introducing a novel perspective on enhancing variational inference methods. The paper addresses a critical limitation in traditional variational autoencoders (VAEs) related to the training criterion being intractable when input noise is introduced.

Key Contributions

The primary contribution of this work lies in the proposal of a modified objective function, termed the "denoising variational lower bound," which is pivotal when the input data is corrupted with noise. This new objective not only remains tractable but also empirically shows an improved performance over conventional VAEs and importance weighted autoencoders (IWAEs). The denoising variant of the variational autoencoder (DVAE) contributes to a significantly better log-likelihood on datasets like MNIST and Frey Face, evidencing the merit of incorporating noise at both input and latent levels.

Technical Overview

Traditional denoising autoencoders (DAEs) incorporate noise injection at the input level, whereas the VAE framework introduces noise into the stochastic hidden layer. Recognizing the potential benefits of simultaneous noise dealings, the authors advocate for noise injection at both stages. The standard VAE lower bound becomes intractable under input corruption due to the requirement of marginalizing over the input noise. This paper circumvents the intractability by proposing a practical, modified bound that allows for efficient training even when noise is present in the input data.

For variational inference tasks, the paper employs a comprehensive theoretical analysis, detailing the construction of a new class of approximate distributions that involve marginalizing input noise over a corruption distribution, potentially achieving a mixture of Gaussian distributions. This formulation allows for enhanced flexibility in modeling complex problem domains that challenge simple VAE structures.

Experimental Validation

The experimental results substantiate the core claims, with empirical evaluations on the MNIST and Frey Face datasets demonstrating the DVAE's superiority over baselines without input noise incorporation. The improved average log-likelihood scores signal a meaningful stride toward robust variational modeling techniques suitable for real-world applications.

Implications and Future Directions

The implications of this work are both practical and theoretical. Practically, the DVAE framework potentially offers a more resilient alternative for applications demanding robust generalization in the presence of input noise. Theoretically, the introduction of the denoising variational lower bound expands the toolkit available for probabilistic model training, inviting further exploration into optimized corruption strategies and more general noise models.

Future research may pivot toward learning parameterized corruption functions or exploring entirely different forms of noise distributions to enhance VAE performance further. Moreover, there is room to investigate the denoising criterion’s utility across other model architectures or domains beyond those considered in the paper.

This paper contributes a meaningful addition to the field of probabilistic modeling and variational inference, particularly for researchers looking for refined techniques to empower autoencoder frameworks in noisy settings. Such advancements could set the stage for more adaptive and versatile neural network models in complex data environments.