Papers
Topics
Authors
Recent
Search
2000 character limit reached

Likelihood-Free Variational Inference

Updated 8 March 2026
  • LFVI is a class of methods for approximate Bayesian inference that bypass the need for tractable likelihoods by using simulator-generated samples.
  • It employs techniques such as forward KL objectives, variational ABC, synthetic likelihoods, and density-ratio estimation to construct scalable surrogate models.
  • LFVI offers significant speed and scalability advantages while addressing challenges like gradient variance and simulator efficiency in complex models.

Likelihood-Free Variational Inference (LFVI) refers to a class of methods for approximate Bayesian inference in models where the likelihood is intractable or not explicitly available, but one can generate samples from the model—typically via a simulator. LFVI generalizes classical variational inference to permit black-box simulators, implicit models, and scenarios common in simulation-based inference or scientific modeling, where densities are unavailable but simulation is straightforward. The methodology encompasses a range of approaches including forward KL-based objectives, variational ABC, synthetic likelihoods, density-ratio estimation, and divergence-minimization frameworks.

1. Key Principles and Theoretical Foundations

LFVI replaces the requirement to evaluate the likelihood in the variational objective with alternatives, including sample-based surrogates, scoring rules, or estimators that depend only on simulator draws. This enables inference in problems where standard variational techniques are rendered infeasible by intractable or missing likelihoods.

LFVI formulations may utilize:

  • Forward KL variational loss: Minimizing KL[p(x,z)q(x,z)]\mathrm{KL}[p(x,z) \| q(x,z)] leads to a loss functional

LFA[q]=p(x,z)logq(zx)dxdz\mathcal{L}_{FA}[q] = -\iint p(x, z) \log q(z|x) dx dz

which is minimized by the exact posterior marginals under mean-field factorization and is purely likelihood-free, requiring only simulated pairs (x,z)(x,z) (Ambrogioni et al., 2018).

  • Variational objectives for ABC: The ELBO is constructed for the ABC posterior, with reparameterization and automatic differentiation yielding low-variance stochastic gradients (Moreno et al., 2016).
  • Density-ratio estimation: The intractable likelihood ratio logp(x,zβ)logq(x,z)\log p(x,z|\beta) - \log q(x,z) is replaced by a learned surrogate r(x,z,β;θ)r(x,z,\beta;\theta), typically trained via a proper scoring rule (e.g., logistic loss), enabling surrogate ELBO optimization (Tran et al., 2017).
  • Synthetic likelihood: When summary statistics S(y)θS(y) \mid \theta are approximately Gaussian, the likelihood is approximated by a parametric (Gaussian) model whose moments are estimated via simulation (Ong et al., 2016).
  • Optimal transport divergences: Wasserstein VI generalizes the variational objective to divergences such as the entropic-regularized OT, with gradients obtained via differentiable Sinkhorn iterations, again requiring only simulated samples (Ambrogioni et al., 2018).

2. Algorithmic Methodologies and Loss Construction

LFVI approaches employ varied strategies to define tractable objectives and compute gradients:

  • Monte Carlo gradient estimators: Gradients are estimated by sampling tuples from the simulator and using stochastic optimization. For forward-KL, the unbiased gradient estimator requires only

θLFA1Ss=1Sθlogqθ(z(s)x(s))\nabla_\theta \mathcal{L}_{FA} \approx - \frac{1}{S} \sum_{s=1}^S \nabla_\theta \log q_\theta(z^{(s)}|x^{(s)})

where (x(s),z(s))p(x,z)(x^{(s)},z^{(s)}) \sim p(x,z) (Ambrogioni et al., 2018).

  • Reparameterization tricks: Both the simulator and variational family are reparameterized into differentiable deterministic functions of random seeds to exploit pathwise gradients and permit automatic differentiation (Moreno et al., 2016).
  • Multilevel Monte Carlo (MLMC) and RQMC: Recent advances provide unbiased gradient estimators for variational objectives when only unbiased likelihood estimates are available, leveraging MLMC telescoping and randomized quasi-Monte Carlo to reduce variance and accelerate convergence (He et al., 2021).
  • Ratio learning: LFVI with implicit models typically trains a "discriminator" or ratio estimator to project the intractable terms in the ELBO, alternating between optimizing the ratio estimator (via a proper scoring rule) and the variational parameters (Tran et al., 2017).
  • Amortized inference: In many settings, variational posteriors are amortized across data via inference networks, enabling scalability and efficient test-time inference (Ambrogioni et al., 2018, Nautiyal et al., 2024).

3. Variants and Representative Frameworks

Several representative formulations have been developed and empirically validated:

Method Key Idea Notable Features
FAVI (Ambrogioni et al., 2018) Forward KL on joint (x,z)(x,z) Exact marginals for mean-field qq; trivial marginalization over nuisance variables; no density or gradient of simulator required
AVABC (Moreno et al., 2016) Variational inference for ABC Simulator and variational family are reparameterized; low-variance gradients via autodiff
VBIL (Tran et al., 2015) VB with unbiased likelihood estimator Works in augmented (θ,z)(\theta,z); provably asymptotic variance grows linearly in noise
VBSL (Ong et al., 2016) VB with synthetic (Gaussian) likelihoods Efficient natural-gradient updates; parametric summary likelihood reduces variance
Wasserstein VI (Ambrogioni et al., 2018) Use of OT divergences, Sinkhorn loss Stable, likelihood-free, differentiable framework for implicit models
EP-ABC (Barthelmé et al., 2011) Expectation Propagation for ABC Local constraints per datapoint, high computational efficiency, summary-free posteriors possible
LFVI for HIMs (Tran et al., 2017) Density ratio estimation (GAN-style) Black-box inference for hierarchical implicit models; implicit qq matches implicit simulator flexibility
VAE-based SBI (Nautiyal et al., 2024) Variational Autoencoders for simulators Encoder-decoder pairs trained on synthetic (θ,y)(\theta, y) pairs; adaptive or standard prior on latent zz

FAVI is particularly notable for its minimal reliance on model details; it can perform variational marginalization exactly under mean-field assumptions, with remarkable applications in time series forecasting and meta-classification (Ambrogioni et al., 2018).

4. Applications and Empirical Performance

LFVI has been successfully applied in diverse domains:

  • Simulator-based inference (SBI): Forward amortized inference allows Bayesian forecasting in chaotic dynamical systems, outperforming extended Kalman filters and recovering full trajectories with correct uncertainty coverage (Ambrogioni et al., 2018).
  • Meta-inference: By marginalizing over large families of models or nuisance variables, LFVI yields amortized meta-classification networks competitive with classical ensemble methods but with no retraining cost per dataset (Ambrogioni et al., 2018).
  • Bayesian inference for implicit models: HIM-LFVI scales to large physical simulators and GANs, handling high-dimensional latent spaces and discrete outputs (Tran et al., 2017).
  • Synthetic likelihoods in genomics/population genetics: VBSL permits fast variational inference in high-dimensional summary space, outperforming ABC and MCMC in computational efficiency (Ong et al., 2016).
  • Image generative modeling: Likelihood-free VAEs (e.g., EnVAE, FEnVAE) overcome likelihood misspecification and decoder limitations to improve generation and reconstruction quality in high-dimensional observation models (Xu et al., 24 Apr 2025).
  • State-space and time-series models: VBIL and MLMC-based methods achieve tractable inference in problems where particle filters are needed for unbiased likelihood estimates (Tran et al., 2015, He et al., 2021).

Crucially, LFVI methods often deliver orders-of-magnitude speedups relative to MCMC or rejection-based ABC, with similar or better posterior accuracy in the appropriate model classes.

5. Advantages, Limitations, and Theoretical Guarantees

Advantages:

Limitations:

  • Simulator requirements: LFVI methods rely on the ability to generate many independent simulations for each parameter θ\theta; if simulation is expensive or data limited, performance may degrade (Ambrogioni et al., 2018).
  • Gradient variance: Although modern reparameterization and MLMC techniques reduce variance, simulator cost and model complexity can yield high-variance stochastic gradients (Moreno et al., 2016, He et al., 2021).
  • Mode coverage vs. mode-seeking: Forward-KL tends to "cover" all posterior modes (mass-covering behavior), potentially producing overly broad variational posteriors, whereas reverse-KL is mode-seeking (Ambrogioni et al., 2018).
  • Density learning limitations: Certain LFVI objectives (e.g., joint-contrastive forward KL) do not support direct learning of simulator parameters, requiring auxiliary objectives for model calibration (Ambrogioni et al., 2018).
  • Model class dependence: Synthetic likelihood-based VI assumes summary statistics are approximately Gaussian; when this is violated, posterior accuracy may suffer (Ong et al., 2016).

Theoretical guarantees are generally in line with stochastic optimization and variational Bayes: as the variational family becomes rich enough and gradient noise is controlled, convergence to a (local) optimum of the surrogate evidence lower bound is expected. For forward-KL, exact mean-field marginals are guaranteed (Ambrogioni et al., 2018); for MLMC-based estimators, unbiasedness of the gradient and tight ELBOs are established under mild regularity conditions (He et al., 2021).

6. Comparison to Classical and Alternative Approaches

LFVI unifies but is distinct from a range of approximate inference techniques:

  • ABC (Approximate Bayesian Computation): LFVI builds on ABC but leverages variational objectives, proper scoring rules, or density-ratio surrogates rather than rejection sampling or kernel density estimation, thus improving both sample efficiency and scalability (Moreno et al., 2016, Barthelmé et al., 2011).
  • Reverse-KL VI: Traditional variational inference via reverse-KL requires tractable likelihoods and can underestimate posterior variance. Forward-KL and LFVI objectives eliminate this constraint and match true marginals under mean-field structure (Ambrogioni et al., 2018).
  • Score-based, synthetic likelihood, and optimal transport methods: LFVI includes all approaches that can rephrase or bypass the intractable likelihood via tractable sample-based surrogates, including scoring rules, F-divergences, and Wasserstein distances (Ambrogioni et al., 2018, Xu et al., 24 Apr 2025, Ong et al., 2016).
  • Adversarial and GAN-based VI: LFVI generalizes adversarial training by integrating density-ratio estimation within a variational framework, avoiding adversarial instability and mode collapse (Tran et al., 2017).

7. Future Directions and Open Challenges

LFVI continues to evolve, driven by the need for amortized, black-box, and robust inference in scientific, engineering, and generative modeling domains:

  • Algorithmic innovation: Further reduction of gradient variance (e.g., via advanced MLMC/RQMC, adaptive simulation allocation) remains a target for highly complex simulators (He et al., 2021).
  • Implicit and structured variational families: Approaches leveraging deep networks for inference and marginalization over large hypothesis spaces are likely to enable ever more flexible posterior approximations (Tran et al., 2017, Nautiyal et al., 2024).
  • Theory and diagnostics: Enhanced methods for quantifying approximation error (especially in multimodal or non-Gaussian settings), and automated diagnostics of variational bias, are needed (Barthelmé et al., 2011).
  • Unified frameworks: There is interest in further unifying f-divergence, OT, proper scoring rules, and density-ratio based VI under a common likelihood-free framework (Ambrogioni et al., 2018, Xu et al., 24 Apr 2025).
  • Application to real-world scientific domains: Empirical validation in domains such as cosmology, neuroscience, epidemiology, and beyond will continue to guide both methodological refinement and theory.

In summary, likelihood-free variational inference constitutes a foundational methodology for Bayesian inference in the simulator era, synthesizing advances from ABC, VI, GANs, and optimal transport to permit fully amortized, scalable, and robust inference wherever simulation is possible but density evaluation is not (Ambrogioni et al., 2018, Tran et al., 2017, Moreno et al., 2016, Xu et al., 24 Apr 2025, He et al., 2021, Nautiyal et al., 2024, Ong et al., 2016, Ambrogioni et al., 2018, Barthelmé et al., 2011, Tran et al., 2015).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Likelihood-Free Variational Inference (LFVI).