Quasi-Optimal Adjustment Set in Causal Inference
- Quasi-optimal adjustment sets are collections of covariates designed to nearly minimize variance in causal effect estimation while ensuring valid identifiability.
- The methodology uses d-separation, enumerates valid adjustment sets, and applies ε or δ tolerance parameters to assess trade-offs in high-dimensional settings.
- Empirical studies demonstrate that quasi-optimal sets achieve up to 20% variance reduction and lower sample complexity compared to complete valid adjustment sets.
A quasi-optimal adjustment set is a collection of covariates employed in covariate adjustment for causal effect estimation that—by construction or design—minimizes (or nearly minimizes) the variance of the estimator among all valid adjustment sets, subject to identifiability and practical considerations. This concept enables practitioners to balance statistical efficiency with the measurement or computational cost, while maintaining identification of the causal estimand in both finite- and high-dimensional or longitudinal settings. The quasi-optimality notion generalizes full optimality by permitting tolerance parameters (additive or multiplicative variance gaps) and approximate adjustment conditions when either finite-sample, computational, or model complexity considerations preclude oracle-optimality.
1. Adjustment Sets and Optimality: Foundational Concepts
A valid adjustment set for estimating the causal effect of a treatment on an outcome in a causal DAG (or more generally, a structured graphical model) must satisfy:
- No element of is a descendant of along a causal path to ;
- blocks (d-separates) all non-causal (back-door) paths from to .
The classic adjustment formula asserts:
Optimal adjustment seeks a valid such that the variance of the causal effect estimator is minimized among all valid adjustment sets. In both linear models and semiparametric settings, this unique set minimizes the asymptotic or PAC (Probably Approximately Correct) error, and, in high-dimensional settings, obviates the exponential sample complexity imposed by unnecessary high-cardinality adjustments (Henckel et al., 2019, Choo et al., 2024, Rotnitzky et al., 2019, Adenyo et al., 2024).
2. Formal Definition and Characterization of Quasi-Optimal Adjustment Sets
Quasi-optimality formalizes near-minimal variance subject to either a multiplicative or additive tolerance. Given , the (oracle) optimal adjustment set, a set is -quasi-optimal if
or -quasi-optimal if
where is the estimator of the mean under intervention adjusting for (Rotnitzky et al., 2019).
In summary causal graphs (SCGs) or high-dimensional settings where the structure is only partially known or data are discrete, the “quasi-optimal” set is the smallest valid adjustment (according to a generalized back-door criterion) that in some compatible fine timescale DAG attains the optimal variance, and in every compatible DAG contains the non-descendant part of the DAG-optimal set (Belciug et al., 20 Dec 2025).
3. Algorithms and Graphical Criteria for Construction
DAG and CPDAG Models
The optimal adjustment set in a DAG is characterized graphically by
where are the mediators from to and are their descendants and itself (Henckel et al., 2019, Rotnitzky et al., 2019).
To construct (quasi-)optimal sets:
- Enumerate all valid sets using back-door and d-separation criteria.
- Evaluate the asymptotic variance for regression, IPW, or AIPW estimators.
- Select a minimal set or those satisfying user-specified quasi-optimality tolerances (Rotnitzky et al., 2019, Luo et al., 2024).
Constraint-Based and PAC Approaches
For high-dimensional and discrete distributions, the ε-Markov blanket approach approximates the adjustment set by identifying satisfying
and bounds the corresponding bias by where is a positivity constant (Choo et al., 2024). Two algorithmic approaches are proposed:
- AMBA: Exhaustively enumerates subsets and selects the smallest passing a conditional-independence test for and given .
- BAMBA: Further shrinks to a set satisfying screening-set conditions involving approximate independence for both treatment and outcome, subject to error bounds.
The combined PAC bound asserts that, with enough samples and careful misspecification control, adjusting on a quasi-optimal subset achieves accuracy close to direct adjustment on the large (oracle) valid set, with drastically reduced sample complexity (Choo et al., 2024).
Time-Dependent/LONGITUDINAL and Abstracted Graphs
For time-dependent interventions or SCGs,
- One exploits new inclusion/exclusion lemmas leveraging d-separation on expanded blocks of variables (“Q-nodes”), constructing a sufficient set via greedy forward/backward moves.
- The quasi-optimal set is the terminal set of such an expansion, provably attaining minimum achievable variance relative to all reachable sufficient sets by such graphical moves (Adenyo et al., 2024).
In SCGs, the quasi-optimal adjustment set is computed by constructing parent sets of extended causal nodes, excluding possible descendants of the treatment, tailored to the specific identifiability scenario (Belciug et al., 20 Dec 2025).
4. Variance Comparison, Theoretical Guarantees, and PAC Bounds
Comparison criteria are graphically characterized. For two valid sets and , if and only if
This yields both constructive pruning procedures and certificates of (quasi-)optimality (Rotnitzky et al., 2019, Henckel et al., 2019).
In finite/high-dimensional settings, PAC bounds quantify estimation error, showing that the sample complexity is exponential in the size of the adjustment used, but can be dramatically reduced by using (quasi-)optimal Markov blankets or screened subsets, at the cost of a controlled bias (Choo et al., 2024).
5. Connections to Minimal, Sufficient, and Exhaustive Sets
Enumeration of all sufficient adjustment sets enables principled selection of quasi-optimal sets via explicit trade-offs:
- Minimal cardinality or sparsity
- Asymptotic variance (or plug-in variance estimate)
- Practical constraints (measurement or computational resource demands)
By enumerating all sufficient sets (e.g., all satisfying ), one can extract all adjustment sets, then select among them those that are quasi-optimal under the stated variance criterion (Luo et al., 2024).
6. Practical Implications and Applied Examples
Empirical results from simulation and real-data studies confirm that quasi-optimal adjustment sets systematically yield lower or comparable standard errors compared to naively larger valid sets, without incurring bias or relying on perfect causal structure recovery. In high-dimensional situations, this may mean using just a sparse subset (e.g., an approximate Markov blanket rather than the full valid set), leading to substantial reductions in sample complexity and estimator variance (Choo et al., 2024). In time-varying or SCG settings, quasi-optimal sets can reduce variance by 10–20% over previous graph-based criteria (Belciug et al., 20 Dec 2025, Adenyo et al., 2024).
| Model Class | Adjustment Set Criterion | Quasi-Optimality Features |
|---|---|---|
| Classical DAGs | Parents of mediators forbids | Minimal variance, proven optimality |
| High-dimensional discrete | -Markov blankets, screening sets | Controlled bias, reduced dimensionality |
| Time-dependent/SCGs | Back-door in summary graph, parent sets | Compatible with all fine-scale DAGs |
7. Theoretical and Methodological Extensions
Quasi-optimal adjustment set theory continues to extend to:
- CPDAGs, maximal PDAGs, and summary graphs, allowing for ambiguity and latent summary structure (Belciug et al., 20 Dec 2025, Henckel et al., 2019);
- Nonparametric models: all key results generalize from linear to nonparametric settings using influence function calculus and plug-in variance estimates (Rotnitzky et al., 2019);
- Exhaustive enumeration and selection-based practical workflows, including open-source code, for large but feasible (Luo et al., 2024);
- Constraint-based and data-driven selection under finite-sample PAC error bounds and misspecification robustness (Choo et al., 2024).
Altogether, the notion of quasi-optimal adjustment sets enables rigorous, sample- and graph-aware selection of adjustment covariates with guarantees for identifiability, efficiency, and scalability in modern causal inference (Choo et al., 2024, Belciug et al., 20 Dec 2025, Rotnitzky et al., 2019, Adenyo et al., 2024, Henckel et al., 2019, Luo et al., 2024).