Dynamic Structural Causal Models
- DSCMs are dynamic extensions of SCMs that model time-indexed, stochastic processes with evolving causal mechanisms.
- They integrate trajectories, differential equations, and VAR models to enable rigorous intervention analysis and equilibrium representation.
- Applications span robotic planning and high-dimensional control, demonstrating robust causal inference and structural modularity.
Dynamic Structural Causal Models (DSCMs) generalize classical Structural Causal Models (SCMs) to encompass temporally evolving systems, where each endogenous variable is a function or stochastic process over time. DSCMs integrate the intervention semantics and graphical representations of SCMs with the dynamics of stochastic processes, ordinary or stochastic differential equations, and time-varying structural equations. This enables principled reasoning about interventions and identification of causal effects in time-dependent, high-dimensional, and potentially cyclic or confounded domains.
1. Formal Definition and Core Constructs
A Dynamic Structural Causal Model formalizes temporal causality by treating each endogenous "variable" as a time-indexed process (discrete or continuous), possibly vector-valued and stochastic. In both discrete and continuous time, a DSCM is defined by:
- A set of endogenous processes and exogenous (noise) processes ;
- A time domain , which may be discrete () or continuous ();
- Domains of the processes, for Skorokhod-space trajectories;
- Causal mechanisms mapping parent process histories and exogenous input histories to each process, with a constraint that at time only depends on parent and noise histories up to time ;
- An exogenous noise law with a product structure.
This framework admits both cycles and bidirected edges (from shared exogenous sources), generalizing the acyclic structure of conventional SCMs. In continuous-time settings, these mechanisms correspond to collections of (possibly coupled) differential or stochastic differential equations, while in discrete time they reduce to systems of difference equations or VAR (vector autoregressive) models (Boeken et al., 3 Jun 2024, Cinquini et al., 14 Oct 2024, Bongers et al., 2018, Rubenstein et al., 2016).
2. Intervention Semantics and Markov Properties
Interventions in DSCMs extend the do-operator to trajectories or finite intervals:
- Hard interventions: Replacing a process or derivative (over all or part of ) by a fixed value or exogenous trajectory.
- Additive and forcing interventions (in discrete time): Shifting a component by a fixed value or forcing it towards a target process, respectively.
Upon intervention, the post-interventional DSCM replaces the mechanism(s) for the intervened variables by constants or functions reflecting the intervention, leaving other mechanisms invariant (Cinquini et al., 14 Oct 2024, Peters et al., 2020).
DSCMs inherit a graphical Markov property: conditional independence relations implied by d-separation (or σ-separation, in the cyclic or latent-confounded case) in the mixed graph of the DSCM reflect conditional independence in the joint law of the processes, conditional on the past or appropriate adjustment sets (Boeken et al., 3 Jun 2024, Bongers et al., 2018).
3. Discrete-Time Systems, VAR Models, and Equilibrium Representation
For discrete-time multivariate systems, such as those modeled by VAR(p) processes,
DSCMs provide a route to equilibrium SCMs. Under stability (all eigenvalues of outside the unit disk) and contractivity of the noise, the process converges in distribution to a unique . At equilibrium, the law of can be represented by a (possibly cyclic and confounded) linear SCM,
with , and derived from the original innovations and the impulse response of the process (Cinquini et al., 14 Oct 2024, Ferreira et al., 2023).
Interventional calculations (additive, forcing) at equilibrium correspond to modifications of or , and causal effects at any forecast horizon can be computed analytically using the impulse response matrices. The pipeline for DSCM inference thus involves estimation (e.g., OLS), equilibrium SCM construction, and causal effect computation via do-calculus (Cinquini et al., 14 Oct 2024, Ferreira et al., 2023).
4. Dynamic Structure, Time-Varying and Nonlinear Models
DSCMs can be extended to accommodate time-varying structures, where the contemporaneous and lagged coefficients evolve as smooth functions of time. In this context:
- Time-varying graphs (with adjacency matrices ) encode causal relations that can drift temporally.
- Basis expansions (e.g., B-splines) are utilized to parameterize and lagged weights , constraining the system to a low-dimensional, smoothly varying subspace.
- Structure learning is formulated as a constrained variational inference problem, augmenting a structural VAE with continuous acyclicity penalization (e.g., via NOTEARS-type constraints).
- The approach enables both forecasting of future system structure and counterfactual or interventional predictions with time-dependent effect estimation (Wang et al., 11 Jan 2025).
Identifiability of the contemporaneous graph at each is ensured under independent non-Gaussian or isotropic Gaussian noise. Dynamic causal contrasts (e.g., time-indexed average treatment effects) can be expressed in closed form and tracked over time (Wang et al., 11 Jan 2025).
5. Continuous-Time Systems, ODEs, and SDEs
Continuous-time DSCMs generalize to systems described by ODEs or SDEs:
- Deterministic kinetic models: Each process evolves according to , with initial values and (optionally) independent measurement noise (Peters et al., 2020).
- Stochastic kinetic models: SDEs of the form capture continuous stochastic perturbations.
- Structural equations in the DSCM correspond to maps from parent trajectories (and initial conditions) to response trajectories; interventions correspond to replacement (surgical) modifications of the relevant drift and/or diffusion terms (Rubenstein et al., 2016, Peters et al., 2020, Boeken et al., 3 Jun 2024).
DSCMs admit time-splitting and subsampling operations, making it possible to relate continuous-time structural models to vector-autoregressive forms at discrete time grids, and to deploy constraint-based discovery algorithms (e.g., FCI) post-discretization (Boeken et al., 3 Jun 2024).
At equilibrium (as ), under stability or attractivity, continuous-time (S)DCMs reduce to classical SCMs, allowing established identification and adjustment criteria to apply (Bongers et al., 2018, Rubenstein et al., 2016).
6. Modularization, Expressivity, and Computational Aspects
Dynamic SCMs achieve modularity and temporal representational efficiency via:
- Socket and buffer variables, enabling dynamic attachment/detachment of agents, modular interaction, and constant-space temporal representation (as in multi-agent robotics and cyber-physical systems).
- Time context variables and buffer monads, allowing efficient threading of time indices without full time-unrolled graphs.
- Relaxed stationarity, permitting agent populations and structure to evolve while preserving local modeling invariances (Howard et al., 3 Jun 2024).
In robotic planning and high-dimensional control applications, DSCM-based learning leverages causal structure distributions, distributing causal and functional uncertainty explicitly across the system. Latent-variable parameterizations (encoder-multidecoder constructions) and distributional inference over graph structure yield robust, sample-efficient forecasting and effective adaptation to interventions or distributional shifts (Murillo-Gonzalez et al., 8 Aug 2025).
7. Identifiability, Graphical Criteria, and Estimation
Identifiability in DSCMs depends on the following:
- Stationarity and stability of the underlying process (VARs, ODE/SDEs).
- Known or estimable block structure (for noise orthogonalization or ICA-based identification in the presence of hidden confounders).
- For time-varying causal graphs, identifiability of the DAG at each timepoint with independent non-Gaussian or isotropic Gaussian errors (Cinquini et al., 14 Oct 2024, Wang et al., 11 Jan 2025, Ferreira et al., 2023).
Graphical criteria from the summary or full-time causal graph (SCG or FTCG) provide systematic conditions for when direct effects are identifiable from observed data, including explicit adjustment set constructions (big and single-door adjusters) and partial regression estimators in linear-Gaussian cases (Ferreira et al., 2023).
DSCMs thus unify and extend core causal identification principles—back-door, front-door, and more general ID algorithms—to time-indexed, possibly cyclic and multi-agent domains.
References:
- (Cinquini et al., 14 Oct 2024) A Practical Approach to Causal Inference over Time
- (Peters et al., 2020) Causal models for dynamical systems
- (Wang et al., 11 Jan 2025) Dynamic Causal Structure Discovery and Causal Effect Estimation
- (Bongers et al., 2018) Causal Modeling of Dynamical Systems
- (Rubenstein et al., 2016) From Deterministic ODEs to Dynamic Structural Causal Models
- (Howard et al., 3 Jun 2024) Extending Structural Causal Models for Autonomous Vehicles to Simplify Temporal System Construction & Enable Dynamic Interactions Between Agents
- (Boeken et al., 3 Jun 2024) Dynamic Structural Causal Models
- (Murillo-Gonzalez et al., 8 Aug 2025) Learning Causal Structure Distributions for Robust Planning
- (Ferreira et al., 2023) Identifiability of Direct Effects from Summary Causal Graphs