Variational Foresight Dynamic Selection (VFDS)
- VFDS is a Bayesian framework for time-varying, context-dependent variable selection that uses variational inference for scalable posterior and predictive analysis.
- It dynamically models predictors' relevance in high-dimensional regression, yielding interpretable sparsity patterns and improved forecasting in economics and finance.
- The algorithm employs coordinate-ascent variational Bayes with Polya–Gamma augmentation, achieving linear computational complexity and robust handling of abrupt shifts in predictor importance.
Variational Foresight Dynamic Selection (VFDS) is a Bayesian framework for time-varying, context-dependent variable selection in high-dimensional dynamic models. VFDS is designed to infer dynamically evolving predictive structures, such as changes in which input features or predictors are relevant at each time point, while maintaining computational scalability through variational methods. The framework achieves efficient posterior and predictive inference in time-varying parameter (TVP) regression models, offering robust dynamic variable selection and interpretable sparsity patterns foundational for forecasting in economics, finance, and other domains characterized by temporally dependent data (Koop et al., 2018, Bianco et al., 2023).
1. Problem Formulation and Model Structure
VFDS addresses the challenge where predictor relevance changes over time and across contexts, requiring a mechanism to “foresee” which variables are likely to be informative before observing all data. The model is based on a state-space TVP regression:
with latent coefficients that evolve as:
%%%%2%%%%
where is a time-varying binary inclusion indicator controlled by a Bernoulli–Gaussian (“spike-and-slab”) hierarchy:
The variances and hyperparameters () may themselves be endowed with conjugate inverse-gamma priors (Bianco et al., 2023). This structure enables both smooth time evolution and abrupt switches in variable importance.
The observation noise variance may evolve with log-stochastic volatility:
where is the tridiagonal DLM precision structure encoding temporal regularization.
2. Bayesian Foresight and Dynamic Selection
VFDS operationalizes dynamic selection (“foresight”) by modeling the inclusion probabilities
with the latent process evolving as a Gaussian Markov random field (GMRF). This captures smooth persistence in the probability of each variable being active, while allowing context-aware, rapid switches—pockets of predictability—driven by the data.
In practice, raw inclusion probabilities can be further regularized by spline-based smoothing, minimizing KL divergence from the variational posterior (Bianco et al., 2023). This approach encourages interpretable and stable time-varying sparsity.
3. Variational Inference and Algorithmic Details
Inference is conducted via coordinate-ascent variational Bayes (VB), using a mean-field factorization:
where are Polya–Gamma auxiliaries introduced for tractability in the Bernoulli–logit link. The evidence lower bound (ELBO) is optimized:
Key closed-form updates include:
- : multivariate Gaussian, involving inversion of tridiagonal precision matrices in time per ;
- : Bernoulli with mode , where
- : multivariate Gaussian, with blockwise structure;
- : Polya–Gamma, facilitating efficient updates to the logit inclusion link;
- : handled via specialized Newton-type algorithms for non-conjugate log-linear variance (Bianco et al., 2023).
Early dropping of variables (when remains below a threshold) further accelerates computation.
Algorithmic summary (CAVI framework):
- Initialize all means and covariances.
- Iterate:
- For each variable , update .
- For each time , update .
- Update (or if variance is homoskedastic).
- Stop on convergence of the ELBO or a chosen tolerance.
Computational complexity is per iteration, supporting applications with –$400$ and up to several hundreds (Bianco et al., 2023, Koop et al., 2018).
4. Forecasting, Foresight, and Predictive Distributions
Posterior-based foresight is central, yielding one- or multi-step-ahead predictions that reflect both coefficient uncertainty and the dynamic sparsity pattern:
Predictive variance includes both coefficient and selection uncertainty:
Monte Carlo samples can be drawn from the full variational posterior for density forecasting, log-score computation, and uncertainty quantification (Bianco et al., 2023, Koop et al., 2018).
5. Empirical Performance and Scalability
Extensive simulation studies demonstrate that VFDS outperforms both static (horseshoe, SSVS, EMVS, etc.) and dynamic (DVS, DSS) Bayesian selection methods on synthetic and real data:
- Simulation: Posterior overlap of with MCMC posterior for always-active parameters is 80–90%; for inactive parameters ≈75%. Lowest mean squared error (MSE) among all competitors. High F1-scores for time-localized inclusion patterns.
- Macroeconomic forecasting: In FRED-QD and other large macroeconomic datasets (), VFDS achieves point and density forecast improvements over unobserved-component and rolling-TVAR models for horizons up to quarters. Predictors selected exhibit interpretable temporal dynamics, revealing time-varying Phillips-curve and business-cycle information.
- Financial forecasting: In equity premium prediction (), persistent selection of a small set of variables (e.g., “max-return,” turnover) aligns with known economic mechanisms, with superior log-score and MSE relative to static/rolling competitors.
Time complexity is linear in both dimensions: per iteration. Empirically, VFDS converges in a few dozen iterations and is $3$– faster than prior dynamic spike-and-slab variational Bayes and over an order of magnitude faster than MCMC (Bianco et al., 2023).
6. Interpretability and Application Domains
The temporal trajectories of the inclusion probabilities provide interpretable maps of variable relevance over time, supporting scientific and economic hypothesis testing. In inflation forecasting, VFDS selects predictors such as lagged inflation, industrial production, consumer spending, producer price indices, and unemployment—quantifying their dynamic relationship to the forecast target and revealing underlying structures such as demand-supply shocks and shifting Phillips-curve behavior (Bianco et al., 2023). In financial domains, sparsity persists on select portfolios, providing insights into market frictions.
VFDS has also been adapted to domains with cost-aware selection, such as human activity recognition, where the sequential cost-benefit analysis of feature acquisition is crucial for deploying sensor-based systems (Ardywibowo et al., 2022).
7. Summary Table: Core Elements of VFDS
| Component | Description | Core Reference |
|---|---|---|
| Model Structure | TVP regression + dynamic spike-and-slab + stochastic volatility | (Koop et al., 2018) |
| Inference | Coordinate-ascent mean-field variational Bayes with Polya–Gamma augm. | (Bianco et al., 2023) |
| Dynamic Selection Mechanism | GMRF-evolving logit → time-varying | (Bianco et al., 2023) |
| Computational Complexity | per VB iteration (with drop rules) | (Bianco et al., 2023) |
| Predictive Capability | Superior out-of-sample point/density forecast, interpretable sparsity | (Bianco et al., 2023) |
VFDS is a general, computationally efficient strategy for dynamically and adaptively selecting predictors, suited for high-dimensional settings and supporting interpretable inference on time-varying structures. Its design principles and algorithmic underpinnings position it as a robust foundation for foresight-aware modeling in temporally structured, high-dimensional regression and forecasting applications (Bianco et al., 2023, Koop et al., 2018).