Identifiability of latent causal graphical models without pure children (2505.18410v1)
Abstract: This paper considers a challenging problem of identifying a causal graphical model under the presence of latent variables. While various identifiability conditions have been proposed in the literature, they often require multiple pure children per latent variable or restrictions on the latent causal graph. Furthermore, it is common for all observed variables to exhibit the same modality. Consequently, the existing identifiability conditions are often too stringent for complex real-world data. We consider a general nonparametric measurement model with arbitrary observed variable types and binary latent variables, and propose a double triangular graphical condition that guarantees identifiability of the entire causal graphical model. The proposed condition significantly relaxes the popular pure children condition. We also establish necessary conditions for identifiability and provide valuable insights into fundamental limits of identifiability. Simulation studies verify that latent structures satisfying our conditions can be accurately estimated from data.
Summary
- The paper introduces novel sufficient conditions for model identifiability using a double triangular structure in the Gamma matrix.
- It establishes necessary criteria ensuring that latent causal models are uniquely determined even without the pure children requirement.
- Simulations validate that the proposed framework accurately recovers latent structures, enhancing causal inference in diverse data settings.
This paper addresses the challenging problem of identifying causal graphical models in the presence of unobserved (latent) variables, a crucial task in fields like psychology, education, and machine learning. While existing methods often impose stringent requirements, such as demanding multiple "pure children" (observed variables with only one latent parent) for each latent variable or restricting the latent causal graph structure, this work proposes significantly weaker conditions for model identifiability.
The core contributions of the paper are twofold:
- It introduces novel sufficient conditions for the identifiability of the entire causal model, including the number of latent variables (K), the latent causal graph (Λ), the latent-to-observed bipartite graph (Γ), and the associated probability distributions. These conditions are centered around a "double triangular" structure in the Γ matrix.
- It establishes necessary conditions for identifiability, providing insights into the fundamental limits of learning such models.
The proposed framework operates under a Binary Latent Causal Model (BLCM), which makes the following key assumptions:
- Latent Variables: Assumed to be binary (Hk∈{0,1}).
- Observed Variables: Can be of arbitrary types (discrete, continuous, mixed) from non-degenerate separable metric spaces.
- Assumption 1 (Measurement Model): There are no direct causal edges between observed variables, and no edges from observed to latent variables. Observed variables are thus noisy measurements of the latent variables.
- Assumption 2 (Basic Graphical Model Assumptions): The joint distribution P(X,H) satisfies the causal Markov property with respect to the true DAG G=(X∪H,E) and is faithful to G.
- Assumption 3 (Nondegeneracy):
- All 2K configurations of latent variables have non-zero probability (P(H=h)>0).
- For each observed variable Xj, its conditional distribution P(Xj∣PaG(Xj)=hp) is distinct for every unique configuration hp of its latent parents.
- Every latent variable Hk has at least one observed child (i.e., no all-zero columns in Γ).
The paper uses a strong statistical notion of identifiability, meaning that if two sets of parameters produce the same observed data distribution P(X), then the parameters themselves must be equivalent (up to known ambiguities like label permutation of latents, sign-flipping of binary latent states, and Markov equivalence of the latent graph Λ).
Sufficient Conditions for Identifiability
The central sufficient condition is the double triangular Γ-matrix:
- Definition 3 (Triangular Γ-matrix): A K×K binary matrix Γ1 (a sub-matrix of Γ) is "triangular" if, after some row and column permutations, it becomes a lower (or upper) triangular matrix with ones on the diagonal.
where1 2 3 4 5
Γ₁ = ( 1 0 0 ... 0 ) ( * 1 0 ... 0 ) ( * * 1 ... 0 ) ( : : : ... : ) ( * * * ... 1 )
*
can be 0 or 1. - Definition (Double Triangular): The J×K latent-to-observed adjacency matrix Γ is "double triangular" if its rows can be permuted such that it contains two disjoint K×K triangular sub-matrices, Γ1 and Γ2.
where Γ1,Γ2 are K×K triangular, and Γ3 is the remaining (J−2K)×K part. This condition significantly relaxes the "two pure children per latent" requirement (which corresponds to Γ1=Γ2=IK).1 2 3
Γ = ( Γ₁ ) ( Γ₂ ) ( Γ₃ )
Key Theorems (Sufficient Conditions):
- Theorem 1 (Identifying K): If the true Γ-matrix is double triangular, the number of latent variables K is identifiable.
- Proof Idea: The double triangular structure ensures (via Lemma 1) that certain marginal probability tables involving observed variables (e.g., P(XS1,XS2) where XS1,XS2 correspond to rows of Γ1,Γ2) have a rank of 2K. An alternative model with K′<K latents would yield a rank of 2K′, leading to a contradiction.
- Definition 4 (Subset Condition): The Γ-matrix satisfies the subset condition if for any two distinct latent variables Hk,Hl, the set of observed children of Hk, ChX(Hk), is not a subset of ChX(Hl), and vice-versa. This means no column in Γ is "dominated" by another.
- Theorem 2 (Identifiability of Model Components): Given a known K:
- Proof Idea: Involves a three-way tensor decomposition of the observed data distribution P(XS1,XS2,XS3). Kruskal's theorem guarantees the uniqueness of this decomposition under the full rank conditions provided by Lemma 1 (due to Γ1,Γ2 being triangular) and distinctness of columns in P(XS3∣H) (due to non-empty Γ3 columns and nondegeneracy). The subset condition helps resolve ambiguities in mapping tensor components back to specific latent variable configurations.
The paper highlights that these results establish identifiability against any alternative model satisfying the basic BLCM assumptions, not just alternatives that also meet the double triangular condition. This is a stronger claim than "recoverability" often found in causal discovery literature.
Necessary Conditions for Identifiability
The paper also establishes conditions that must hold for identifiability:
- Theorem 3 (Three Measurements per Latent): For a BLCM with known K and categorical observed variables to be identifiable, each latent variable Hk must have at least three observed children. This is stricter than the often-assumed two measurements.
- Theorem 4 (Subset Condition): For a BLCM with known K≥2 and a known Γ-matrix to be identifiable, Γ must satisfy the subset condition (Definition 4). This is crucial for disentangling latent variable influences and resolving sign-flipping ambiguities.
Counterexamples are provided to illustrate why models violating these necessary conditions (or the nondegeneracy assumptions) become non-identifiable.
Experimental Validation
Simulations were conducted to empirically verify the theoretical identifiability results.
- Setup: A BLCM with K=3 latent variables and J=8 observed variables (structure from Figure 1, which satisfies the double triangular condition). Observed variables included binary and continuous (Normal, Cauchy) types. Different latent graph structures (Λ) were tested: linear, collider, and fully dependent.
- Method: A score-based estimator was used. Continuous responses were dichotomized, and a penalized EM algorithm (adapted from Ma et al., 2023) was employed to estimate Γ and P(H). Λ was then learned from the estimated P(H) using Greedy Equivalence Search (GES).
- Results: The estimated Γ and Λ (measured by Structured Hamming Distance, SHD) showed good accuracy, which improved with increasing sample sizes (N=1000,5000,10000). This supports the claim that complex latent structures satisfying the proposed weak conditions can be accurately recovered from data.
Comparison with Related Work (from Supplement S.1)
- Kivva et al. (2021): Allows arbitrary categorical latents but relies on a strong "mixture oracle" assumption (to determine the number of mixture components in marginals), which may not hold in the paper's general nonparametric setting. Their tensor decomposition also differs, using log number of mixture components.
- Chen et al. (2024): Handles polytomous latents but requires three pure children per latent, stricter cardinality constraints on observed vs. latent variables, and stronger full-rankness nondegeneracy.
- Lee et al. (2025): Focuses on binary latents with parametric GLM assumptions for P(Xj∣H) and typically requires two pure children for strict identifiability, or achieve generic identifiability under GLMs without pure children. This paper offers nonparametric identifiability with weaker structural assumptions (double triangular instead of pure children).
Practical Implications and Implementation
The paper's theoretical results have significant practical implications:
- Broader Applicability: The relaxed conditions allow for the identification of more complex and realistic causal models where latents may not have pure children and observed data is multi-modal.
- Educational Assessment: As an example, student skills (binary latent) can be identified from diverse test responses (observed) even if questions test multiple skills.
- Algorithm Development: The identifiability proofs (especially the tensor decomposition aspect) can guide the development of new learning algorithms. While the paper uses a penalized EM approach for experiments, the theoretical framework using Kruskal's theorem suggests other potential algorithmic avenues.
- Implementation Considerations:
- Verifying Double Triangular Condition: For a given K, this involves searching for specific sub-structures in Γ. This might be computationally intensive for large K,J but can be approached greedily.
- Tensor Decomposition: Practical application of Kruskal's theorem requires robust tensor decomposition algorithms, which can be challenging with finite noisy data. The experimental algorithm sidesteps direct tensor decomposition for estimation.
- Estimating K: While Theorem 1 shows K is identifiable, practical estimation of K often relies on model selection criteria (BIC, etc.) or domain knowledge.
- Computational Requirements: The penalized EM algorithm used in experiments is iterative. Complexity will depend on N,J,K, and the number of EM iterations.
Discussion and Future Work
The authors suggest several avenues for future research:
- Extending beyond measurement models to include direct causal links between observed variables.
- Generalizing identifiability conditions to other types of latent variables (e.g., general categorical, continuous, or mixed).
- Developing more sophisticated nonparametric causal discovery algorithms specifically tailored to these weaker identifiability guarantees, potentially leveraging the full information in continuous responses without dichotomization.
In summary, this paper makes a valuable contribution by significantly weakening the conditions required for identifying latent variable causal models with binary latents. The proposed "double triangular" condition on the latent-to-observed graph, combined with standard assumptions, provides a theoretical foundation for learning more complex causal structures from diverse types of observed data, opening up new possibilities for causal inference in various application domains.