Amortised Variational Inference
- Amortised Variational Inference is a technique that replaces per-instance optimization with a global neural network to perform efficient approximate Bayesian inference in latent variable models.
- It leverages the evidence lower bound (ELBO) to jointly train inference and generative networks, enhancing scalability and enabling single forward-pass inference for unseen data.
- AVI finds applications in deep generative modeling, structured Bayesian inference, and meta-learning, offering practical benefits for fast, scalable probabilistic reasoning.
Amortised Variational Inference (AVI) is a core methodology in contemporary probabilistic machine learning that enables efficient, scalable, and flexible approximate inference in complex latent variable models. By replacing instance-specific optimization of variational parameters with a shared parametric inference mechanism—typically a neural network—AVI is foundational to deep generative modeling, structured Bayesian inference, and probabilistic meta-learning. Recent advances address its theoretical guarantees, computational trade-offs, and algorithmic extensions, with methodological and empirical progress spanning hierarchical Bayesian models, dynamical systems, meta-learning, and stochastic processes.
1. Conceptual Foundation and Mathematical Framework
Amortised Variational Inference seeks to approximate intractable posterior distributions in latent variable models by introducing a parameterized family of tractable distributions and optimizing the reverse Kullback–Leibler divergence via maximization of the evidence lower bound (ELBO): Instead of instantiating and optimising local parameters for each (“classical” mean-field VI), AVI employs a global inference network (encoder) , typically a neural network mapping to the parameters of (Ganguly et al., 2022). This network is trained jointly with the generative model such that, for unseen , variational inference can be executed in a single forward pass with no per-instance optimization.
The ELBO, optimized with stochastic re