Papers
Topics
Authors
Recent
Search
2000 character limit reached

BayesSimIG: Scalable Likelihood-Free Inference

Updated 23 February 2026
  • BayesSimIG is a family of scalable algorithms for likelihood-free Bayesian inference that bypasses intractable likelihoods using simulation-based methods.
  • It employs GPU-accelerated neural posterior estimation, surrogate Gaussian process emulators, and Hamiltonian Monte Carlo to efficiently explore high-dimensional, multimodal posteriors.
  • Empirical results demonstrate significant speed-ups—up to 64×—and practical applicability in robotics, reinforcement learning, and complex time-series modeling.

BayesSimIG is a term describing a family of scalable Bayesian simulation algorithms and software frameworks for likelihood-free inference of model parameters when the likelihood function is difficult or impossible to compute directly. It has been instantiated in multiple domains, most notably for adaptive domain randomization in reinforcement learning with fast GPU simulation (Antonova et al., 2021), indirect inference for models with intractable normalizing functions (Park, 2020), and Bayesian estimation in the context of integer-valued time series models such as geometric INGARCH (Andrews et al., 2024).

1. Overview and Motivation

BayesSimIG addresses the challenge of inferring high-dimensional and often multimodal posteriors over model parameters θ\theta in settings where:

  • The simulator (or model) is treated as a black box, with no tractable likelihood.
  • Only the ability to simulate synthetic data τs\tau^s given parameter settings θ\theta is required.
  • Real data τr\tau^r or sufficient statistics SxS_x are available, but the mapping p(τθ)p(\tau | \theta) is unavailable.

Applications include identifying accurate simulation parameterizations in robotics for sim-to-real transfer (Antonova et al., 2021), statistical network modeling with doubly-intractable distributions (Park, 2020), and estimation and forecasting in state-dependent count time series models (Andrews et al., 2024).

The principal goal is to estimate the posterior p(θsr)p(\theta | s^r), where srs^r is a summary (potentially high-dimensional) of observed data, using only the capability to sample from the simulator.

2. Likelihood-Free Bayesian Inference Framework

The BayesSimIG approach is based on likelihood-free inference, leveraging either neural density estimation (Antonova et al., 2021) or surrogate Gaussian process emulation (Park, 2020), and can be summarized as follows:

  • Simulation-based inference: For sampled θi\theta_i from a proposal or prior p~(θ)\tilde p(\theta), simulate trajectories τis\tau^s_i. Time-series, physical, or count data can be processed.
  • Summary statistics: Each trajectory τis\tau^s_i is mapped deterministically to a summary si=f(τis)s_i = f(\tau^s_i), reducing the trajectory to a lower-dimensional representation.
  • Posterior estimation: A conditional density estimator qϕ(θs)q_\phi(\theta | s) is trained to approximate p(θs)p(\theta | s). Two main instantiations are:

The approximate posterior is given by:

p^(θsr)[p(θ)p~(θ)]qϕ(θsr)\hat{p}(\theta | s^r) \propto \left[ \frac{p(\theta)}{\tilde{p}(\theta)} \right] q_\phi(\theta | s^r)

If the proposal and prior are identical, the ratio cancels.

Iterative schemes can adapt the proposal p~\tilde{p} towards the current posterior, improving efficiency and coverage in high-dimensional problems.

3. Algorithmic and Computational Details

3.1 GPU-Accelerated Neural Posterior Estimation

In the context of large-scale robotics and RL (Antonova et al., 2021):

  • Simulation of 10,000–20,000 instances in parallel is executed on a single NVIDIA A100 GPU using IsaacGym.
  • Summaries include cross-correlation, start, waypoints, and path-signature-based features, implemented in PyTorch or Signatory.
  • Conditional density estimation via MDNN/MDRFF is end-to-end GPU-based.
  • Data flow is pipelined with YAML configuration, batch simulation, summarization, and neural density fitting, all integrated with TensorBoard for diagnostic and posterior visualization.

3.2 Surrogate Gaussian Process Emulator for Intractable Normalization

For doubly-intractable likelihood models (Park, 2020):

  • A surrogate q(Syψ)q(S_y | \psi), where SyS_y is the summary statistic, with ψ=(μ,Σ)\psi = (\mu, \Sigma).
  • Gaussian process prior on the map θμ(θ)\theta \mapsto \mu(\theta):

    μ(θ)GP(m(θ),k(θ,θ))\mu(\theta) \sim \mathcal{GP}(m(\theta), k(\theta,\theta'))

    with Matérn covariance and linear mean.

  • Posterior MCMC proceeds via a surrogate-driven auxiliary variable approach, where statistics drawn from the surrogate replace costly inner simulations.

Empirical results show 2–64×\times reductions in compute time for challenging models, with negligible loss in posterior quality (Park, 2020).

3.3 Hamiltonian Monte Carlo for Time Series Models

For geometric INGARCH modeling (Andrews et al., 2024):

  • Posterior over (ω,α,β,ϕ)(\omega, \alpha, \beta, \phi) is explored with HMC.
  • Priors are set on transformed parameters (log/ logit), and a joint log-posterior is constructed combining the conditional likelihood and prior contributions.
  • Gradients are computed analytically for use in the leapfrog integrator.
  • Predictive inference is performed by sampling forward using posterior draws.

4. Software Architecture and Usage

4.1 Modular Components and Interfaces

BayesSimIG is designed to be modular:

  • Supports various policies (random, RL-based, fixed, or user-provided) for simulation control (Antonova et al., 2021).
  • Summarizers are extensible via subclassing (e.g., BaseSummarizer in bayes_sim_ig).
  • Density estimators may be replaced with alternative normalizing flows or autoregressive models.
  • TensorBoard integration provides scalable monitoring and posterior visualization.

4.2 Practical Workflow

A typical usage scenario—e.g., for robotics or RL—follows this sequence:

  1. Specify model/task, prior, policy, summarizer, neural net architecture in YAML.
  2. Launch 10K+ parallel IsaacGym envs, sample θip~\theta_i \sim \tilde{p}, simulate, summarize.
  3. Train qϕ(θs)q_\phi(\theta | s) on GPU batches.
  4. Evaluate and visualize posteriors via TensorBoard.
  5. Plug posterior samples back into the simulation/RL pipeline for domain-randomized policy training.

A representative Python API:

1
2
3
4
5
from bayes_sim_ig import BayesSimIG

bsim = BayesSimIG(config_path="pendulum_config.yaml", log_dir="logs/pendulum")
bsim.run()  # runs inference and RL for configured iterations
theta_samples = bsim.posterior.sample(1000)
(Antonova et al., 2021).

5. Performance, Scalability, and Empirical Findings

BayesSimIG achieves substantial speed-ups and scalability:

  • In RL/robotics tasks (e.g., D100D \approx 100 parameters, N=10,000N=10,000 per iteration), total runtime is $20$–$30$ minutes per posterior on a single GPU, compared to several hours for CPU-based BayesSim (Antonova et al., 2021).
  • Indirect auxiliary variable MCMC with surrogate GP emulators (IAVM) achieves speedups up to 64×64\times relative to prior DMH methods on network data (Park, 2020).
  • For time series, the HMC-based BayesSimIG yields effective posterior approximate sampling and enables Bayesian predictive forecasting (Andrews et al., 2024).

Best practices include:

  • For dimensionality DD, number of simulations per iteration N1000DN \gtrsim 1000 D if feasible.
  • Use cross-correlation-based summarizers for dynamic systems, switch to simpler ones for high-dimensional θ\theta.
  • Monitor GPU memory in extreme simulation regimes and batch accordingly.
  • Posterior coverage quality is sensitive to proposal adaptation; incorrect adaptation can bias inference.

6. Limitations, Extensions, and Customization

Limitations include:

  • Posterior approximation quality is bounded by the richness and coverage of θ\theta in simulation.
  • GPU memory becomes a constraint for extremely large NN and DD.
  • Online update of the simulation proposal may require retraining neural estimators to avoid proposal-induced bias.

Extensions are readily supported:

  • Custom summarizers and density estimators can be implemented, with specific hooks for PyTorch- and Python-based architectures.
  • External RL frameworks (e.g., Stable Baselines3, RLlib) can be integrated by replacing or extending the training module.
  • BayesSimIG is open source, supporting flexible experimentation and research in domain randomization, likelihood-free inference, and model-based forecasting.

By combining efficient GPU-accelerated simulation, modular neural-density estimation, and adaptive proposal strategies, BayesSimIG provides a scalable and extensible framework for Bayesian parameter inference across stochastic simulation, complex time-series, and reinforcement learning tasks (Antonova et al., 2021, Park, 2020, Andrews et al., 2024).

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to BayesSimIG.