2000 character limit reached
Generalization Gap in Amortized Inference (2205.11640v2)
Published 23 May 2022 in stat.ML and cs.LG
Abstract: The ability of likelihood-based probabilistic models to generalize to unseen data is central to many machine learning applications such as lossless compression. In this work, we study the generalization of a popular class of probabilistic model - the Variational Auto-Encoder (VAE). We discuss the two generalization gaps that affect VAEs and show that overfitting is usually dominated by amortized inference. Based on this observation, we propose a new training objective that improves the generalization of amortized inference. We demonstrate how our method can improve performance in the context of image modeling and lossless compression.
- Mingtian Zhang (22 papers)
- Peter Hayes (8 papers)
- David Barber (54 papers)