Papers
Topics
Authors
Recent
2000 character limit reached

DCL-DECOR: Modular Causal Discovery

Updated 2 January 2026
  • DCL-DECOR is a modular precision-based causal discovery framework that decomposes sample precision matrices to isolate both pervasive and localized latent confounders in linear Gaussian systems.
  • It applies a two-stage strategy using latent-variable graphical lasso and correlated-noise DAG learning to accurately recover bow-free directed acyclic graphs.
  • Empirical evaluations show improved directed-edge F1 scores and reduced structural errors compared to alternative methods under simulated mixed confounding scenarios.

DCL-DECOR refers to a modular, precision-based framework for causal discovery in the presence of mixed latent confounding—where both pervasive (broad) and localized (subset-specific) unobserved confounders affect observed data—within linear Gaussian systems. DCL-DECOR decomposes the observed sample precision matrix into a structured component and a low-rank component, isolates the impact of pervasive confounders, and recovers the directed acyclic graph (DAG) structure using correlated-noise SEM learning, followed by a reconciliation step to ensure bow-freeness. The method establishes identifiability conditions and demonstrates empirical gains in directed-edge discovery under simulated mixed confounding scenarios (Asiaee et al., 31 Dec 2025).

1. Mixed Latent Confounding in SEMs

DCL-DECOR addresses the challenge of causal discovery when observed variables xRpx \in \mathbb{R}^p are generated by a structural equation model (SEM) with both global and local latent confounders:

x=Bx+ε,x = B^\top x + \varepsilon,

where the exogenous noise vector ε\varepsilon is modeled as

ε=Ww+Vv+Uu,\varepsilon = Ww + Vv + Uu,

with wN(0,Ip)w \sim \mathcal{N}(0,I_p) (idiosyncratic noise), vN(0,IrS)v \sim \mathcal{N}(0, I_{r_S}) (localized confounders), and uN(0,IrL)u \sim \mathcal{N}(0, I_{r_L}) (pervasive confounders). VRp×rSV \in \mathbb{R}^{p \times r_S} is column-sparse, representing localized effects; URp×rLU \in \mathbb{R}^{p \times r_L} is dense, modeling widespread confounding; WW is diagonal. The observed covariance Σ\Sigma and precision Θ=Σ1\Theta = \Sigma^{-1} are related by

Σ=TΩT1,Θ=TΩ1T,\Sigma = T^{-\top} \Omega T^{-1}, \quad \Theta = T \Omega^{-1} T^\top,

where T=IBT = I - B and Ω=Var(ε)\Omega = \mathrm{Var}(\varepsilon).

Traditional DAG-learning methods misattribute latent-effects-induced correlations, leading to incorrect causal edge inference. Latent variable graphical models capture undirected structure but fail to orient edges. DCL-DECOR modularizes the problem by first deconfounding for pervasive effects, then learning the directed structure in the conditional model (Asiaee et al., 31 Dec 2025).

2. Precision Decomposition

DCL-DECOR performs a two-stage precision decomposition:

  • At the noise level:

    • Uses the Sherman–Morrison–Woodbury identity to split Ω1\Omega^{-1} into a structured component SεS_{\varepsilon} (from idiosyncratic and localized confounders) and low-rank corrective term LεL_{\varepsilon} (from pervasive confounders):

    Ω1=SεLε,\Omega^{-1} = S_\varepsilon - L_\varepsilon,

    with explicit forms: - Sε=(WW+VV)1=DεCε,S_\varepsilon = (WW^\top + VV^\top)^{-1} = D_\varepsilon - C_\varepsilon, with Dε=(WW)1D_\varepsilon = (WW^\top)^{-1}, Cε=DεVA1VDεC_\varepsilon = D_\varepsilon V A^{-1} V^\top D_\varepsilon, A=I+VDεVA = I + V^\top D_\varepsilon V. - Lε=SεU(I+USεU)1USε.L_\varepsilon = S_\varepsilon U (I + U^\top S_\varepsilon U)^{-1} U^\top S_\varepsilon.

  • At the observed level:

    • By congruence transformation, the observed precision matrix splits as

    Θ=SxLx,\Theta = S_x - L_x,

    where Sx=TSεTS_x = T S_\varepsilon T^\top encodes structured (local, typically sparse) dependencies and Lx=TLεTL_x = T L_\varepsilon T^\top is low-rank, rank rL\leq r_L.

  • Estimation: The decomposition is recovered via convex optimization (“latent-variable graphical lasso”):

minS,L[logdet(SL)+tr(Σ^(SL))+λsRloc(S)+λtr(L)],\min_{S,L} \left[ -\log\det(S-L) + \mathrm{tr}(\widehat{\Sigma}(S-L)) + \lambda_s R_{\mathrm{loc}}(S) + \lambda_* \mathrm{tr}(L) \right],

subject to S0,L0,SL0S \succ 0, L \succeq 0, S-L \succ 0, where RlocR_{\mathrm{loc}} enforces locality/sparsity in SS and the nuclear norm tr(L)\mathrm{tr}(L) proxies low-rank (Asiaee et al., 31 Dec 2025).

3. Correlated-Noise DAG Learning on the Structured Component

Once pervasive confounders have been partialled out, the conditional covariance Σcond=Sx1\Sigma_{\mathrm{cond}} = S_x^{-1} captures only local structure and localized confounders. DCL-DECOR employs a correlated-noise DAG learner (DECOR-GL):

  • Objective:

minB,Sε0L(B,Sε;Σ^cond)\min_{B, S_\varepsilon \succ 0} \,\, \mathcal{L}(B, S_\varepsilon; \widehat{\Sigma}_{\mathrm{cond}})

L=tr[Σ^condTSεT]logdetSε+λBB1+λSSε1,off+ρh(B)\mathcal{L} = \mathrm{tr}[\widehat{\Sigma}_{\mathrm{cond}} T S_\varepsilon T^\top] - \log\det S_\varepsilon + \lambda_B \|B\|_1 + \lambda_S \|S_\varepsilon\|_{1,\mathrm{off}} + \rho h(B)

where h(B)=tr(eBB)ph(B) = \mathrm{tr}(e^{B \circ B}) - p enforces acyclicity, T=IBT = I - B.

  • Optimization: Alternates between graph-step (proximal gradient on BB) and noise-step (graphical lasso on SεS_\varepsilon). After convergence, hard-thresholding is used to obtain a sparse graph (Asiaee et al., 31 Dec 2025).

4. Bow-Freeness Reconciliation

Due to unidentifiability in linear Gaussian ADMGs, "bows"—simultaneous presence of a directed edge and residual correlation between the same variable pair—cannot be inferred. DCL-DECOR enforces bow-freeness post hoc:

  • For each pair with both a directed edge (Bij0B_{ij} \neq 0 or Bji0B_{ji} \neq 0) and bidirected residual entry (Γij0\Gamma_{ij} \neq 0 in Γ=Sε1\Gamma = S_\varepsilon^{-1}), retain the element with larger normalized magnitude; set the other to zero.
  • The bow constant cc (typically c=1c=1) calibrates the comparison.

This ensures the final output is a bow-free mixed graph (Asiaee et al., 31 Dec 2025).

5. Identifiability and Theoretical Guarantees

DCL-DECOR’s structure–low-rank decomposition is uniquely identifiable under the transversality condition (tangent cones at (Sx,Lx)(S_x, L_x) of the structured and low-rank varieties intersect only at the origin) and standard incoherence requirements. Given convergence of the decomposition solver and stability in DAG recovery, the causal target is characterized as the minimal bow-free equivalence class

Ebowmin(Σcond)=argmin{B0+Γoff0:(B,Γ) bow-free,Σ(B,Γ)=Σcond}.\mathcal{E}_{\mathrm{bow}}^{\min}(\Sigma_{\mathrm{cond}}) = \arg\min \{\|B\|_0 + \|\Gamma_{\mathrm{off}}\|_0 : (B,\Gamma) \text{ bow-free}, \Sigma(B,\Gamma) = \Sigma_{\mathrm{cond}} \}.

End-to-end consistency is established: under appropriate conditions, the DCL-DECOR output converges to an element in this equivalence class (Asiaee et al., 31 Dec 2025).

6. Algorithmic Workflow

The DCL-DECOR pipeline follows:

  1. Compute the empirical covariance Σ^\widehat{\Sigma} of observed data.
  2. Solve the structured–low-rank split via convex optimization (latent-variable graphical lasso), yielding (S^x,L^x)(\widehat{S}_x, \widehat{L}_x).
  3. Invert S^x\widehat{S}_x (using sparse Cholesky) to obtain Σ^cond\widehat{\Sigma}_{\mathrm{cond}}.
  4. Apply correlated-noise DAG learning (DECOR-GL) to Σ^cond\widehat{\Sigma}_{\mathrm{cond}}, alternating optimization over BB, SεS_\varepsilon until convergence.
  5. Threshold BB and Γ=Sε1\Gamma = S_\varepsilon^{-1}; enforce bow-freeness using the bow reconciliation rule.
  6. Output: final bow-free estimate (B^,Γ^ε)(\widehat{B}, \widehat{\Gamma}_\varepsilon) (Asiaee et al., 31 Dec 2025).

7. Empirical Evaluation

Extensive synthetic experiments evaluate DCL-DECOR’s performance. Scenarios with p=40p = 40 variables, n=600n = 600 samples, and both pervasive (rLr_L up to 5) and localized (rS=15,r_S=15, support size 6) confounders were considered. DCL-DECOR consistently outperformed alternative methods (DECOR-GL, DeCAMFounder, NOTEARS, GOLEM, GES, LiNGAM) in directed-edge F1F_1 score and structural Hamming distance, especially as the strength and rank of pervasive confounding increased. For example, with qP=3q_P=3, Ud=1.0U_d=1.0, mean F1F_1 was $0.431$ for DCL-DECOR versus $0.266$ for DECOR-GL; mean SHD $55.0$ versus $76.2$. The results demonstrate that decomposing precision matrices to account for mixed confounding and imposing bow-freeness are critical for accurate causal discovery in this regime (Asiaee et al., 31 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to DCL-DECOR.