DCL-DECOR: Modular Causal Discovery
- DCL-DECOR is a modular precision-based causal discovery framework that decomposes sample precision matrices to isolate both pervasive and localized latent confounders in linear Gaussian systems.
- It applies a two-stage strategy using latent-variable graphical lasso and correlated-noise DAG learning to accurately recover bow-free directed acyclic graphs.
- Empirical evaluations show improved directed-edge F1 scores and reduced structural errors compared to alternative methods under simulated mixed confounding scenarios.
DCL-DECOR refers to a modular, precision-based framework for causal discovery in the presence of mixed latent confounding—where both pervasive (broad) and localized (subset-specific) unobserved confounders affect observed data—within linear Gaussian systems. DCL-DECOR decomposes the observed sample precision matrix into a structured component and a low-rank component, isolates the impact of pervasive confounders, and recovers the directed acyclic graph (DAG) structure using correlated-noise SEM learning, followed by a reconciliation step to ensure bow-freeness. The method establishes identifiability conditions and demonstrates empirical gains in directed-edge discovery under simulated mixed confounding scenarios (Asiaee et al., 31 Dec 2025).
1. Mixed Latent Confounding in SEMs
DCL-DECOR addresses the challenge of causal discovery when observed variables are generated by a structural equation model (SEM) with both global and local latent confounders:
where the exogenous noise vector is modeled as
with (idiosyncratic noise), (localized confounders), and (pervasive confounders). is column-sparse, representing localized effects; is dense, modeling widespread confounding; is diagonal. The observed covariance and precision are related by
where and .
Traditional DAG-learning methods misattribute latent-effects-induced correlations, leading to incorrect causal edge inference. Latent variable graphical models capture undirected structure but fail to orient edges. DCL-DECOR modularizes the problem by first deconfounding for pervasive effects, then learning the directed structure in the conditional model (Asiaee et al., 31 Dec 2025).
2. Precision Decomposition
DCL-DECOR performs a two-stage precision decomposition:
- At the noise level:
- Uses the Sherman–Morrison–Woodbury identity to split into a structured component (from idiosyncratic and localized confounders) and low-rank corrective term (from pervasive confounders):
with explicit forms: - with , , . -
- At the observed level:
- By congruence transformation, the observed precision matrix splits as
where encodes structured (local, typically sparse) dependencies and is low-rank, rank .
- Estimation: The decomposition is recovered via convex optimization (“latent-variable graphical lasso”):
subject to , where enforces locality/sparsity in and the nuclear norm proxies low-rank (Asiaee et al., 31 Dec 2025).
3. Correlated-Noise DAG Learning on the Structured Component
Once pervasive confounders have been partialled out, the conditional covariance captures only local structure and localized confounders. DCL-DECOR employs a correlated-noise DAG learner (DECOR-GL):
- Objective:
where enforces acyclicity, .
- Optimization: Alternates between graph-step (proximal gradient on ) and noise-step (graphical lasso on ). After convergence, hard-thresholding is used to obtain a sparse graph (Asiaee et al., 31 Dec 2025).
4. Bow-Freeness Reconciliation
Due to unidentifiability in linear Gaussian ADMGs, "bows"—simultaneous presence of a directed edge and residual correlation between the same variable pair—cannot be inferred. DCL-DECOR enforces bow-freeness post hoc:
- For each pair with both a directed edge ( or ) and bidirected residual entry ( in ), retain the element with larger normalized magnitude; set the other to zero.
- The bow constant (typically ) calibrates the comparison.
This ensures the final output is a bow-free mixed graph (Asiaee et al., 31 Dec 2025).
5. Identifiability and Theoretical Guarantees
DCL-DECOR’s structure–low-rank decomposition is uniquely identifiable under the transversality condition (tangent cones at of the structured and low-rank varieties intersect only at the origin) and standard incoherence requirements. Given convergence of the decomposition solver and stability in DAG recovery, the causal target is characterized as the minimal bow-free equivalence class
End-to-end consistency is established: under appropriate conditions, the DCL-DECOR output converges to an element in this equivalence class (Asiaee et al., 31 Dec 2025).
6. Algorithmic Workflow
The DCL-DECOR pipeline follows:
- Compute the empirical covariance of observed data.
- Solve the structured–low-rank split via convex optimization (latent-variable graphical lasso), yielding .
- Invert (using sparse Cholesky) to obtain .
- Apply correlated-noise DAG learning (DECOR-GL) to , alternating optimization over , until convergence.
- Threshold and ; enforce bow-freeness using the bow reconciliation rule.
- Output: final bow-free estimate (Asiaee et al., 31 Dec 2025).
7. Empirical Evaluation
Extensive synthetic experiments evaluate DCL-DECOR’s performance. Scenarios with variables, samples, and both pervasive ( up to 5) and localized ( support size 6) confounders were considered. DCL-DECOR consistently outperformed alternative methods (DECOR-GL, DeCAMFounder, NOTEARS, GOLEM, GES, LiNGAM) in directed-edge score and structural Hamming distance, especially as the strength and rank of pervasive confounding increased. For example, with , , mean was $0.431$ for DCL-DECOR versus $0.266$ for DECOR-GL; mean SHD $55.0$ versus $76.2$. The results demonstrate that decomposing precision matrices to account for mixed confounding and imposing bow-freeness are critical for accurate causal discovery in this regime (Asiaee et al., 31 Dec 2025).