Pseudo-outcome Imputation in Counterfactual Regression
- The paper introduces a framework (PIPCFR) that integrates post-treatment variables into pseudo-outcome imputation for more accurate individual treatment effect estimation.
- It establishes a novel theoretical risk bound that balances variance reduction and bias control, outperforming prior pseudo-outcome and matching-based methods.
- Empirical evaluations across simulations and real-world datasets demonstrate significant performance gains and robustness under complex post-treatment conditions.
Pseudo-outcome Imputation with Post-treatment Variables for Counterfactual Regression (PIPCFR) is an algorithmic framework for individual treatment effect (ITE) estimation in observational studies that formally integrates post-treatment variables into pseudo-outcome construction. The core innovation is to improve counterfactual outcome prediction by representing and leveraging information in post-treatment variables while actively controlling the bias introduced by their post-treatment status. The approach is rigorously characterized by a novel theoretical risk bound and demonstrates empirical superiority over prior pseudo-outcome and matching-based methods in a variety of simulation and real-world settings (Lin et al., 21 Dec 2025).
1. Problem Setting and Motivation
In the standard Neyman–Rubin potential outcomes framework, each observational unit is described by pre-treatment covariates , a binary treatment indicator , post-treatment variables (observed only under the assigned treatment), and an outcome . The objective is to estimate the ITE for each , but the fundamental challenge is that only one of the two potential outcomes is observed per unit.
Traditional approaches address this via imputation—using models trained on observed data to predict missing counterfactuals (“pseudo-outcomes”). However, most prior work overlooks the role of post-treatment variables, which may contain supplemental information about outcome-relevant stochasticity. The omission of such variables can lead to increased variance in counterfactual predictions. PIPCFR introduces a pseudo-outcome construction mechanism that conditions on both and learned representations of , with techniques to mitigate spurious biases from using post-treatment data (Lin et al., 21 Dec 2025).
2. Theoretical Formulation and Risk Bound
PIPCFR defines a pseudo-outcome constructor , where is a learned representation of post-treatment features. For each instance, serves as a pseudo-label for the unobserved counterfactual. The ITE error is
which cannot be directly minimized due to missing counterfactuals.
PIPCFR establishes the following risk upper bound under standard assumptions (overlap, SUTVA, unconfoundedness):
where:
- is a PIP (pseudo-outcome imputation proxy) loss capturing student-pseudo-outcome and teacher-factual fit,
- quantifies the distance (e.g., via MMD or KL) between representations of under each treatment conditional on (penalizing post-treatment bias),
- is the teacher’s counterfactual error,
- is the squared error between and ,
- is the factual regression error for treatment .
Incorporating enables variance reduction in counterfactual estimates, but also risks bias due to treatment-induced shifts in the distribution of . The IPM/KL term controls this bias by enforcing , trading variance reduction against bias mitigation (Lin et al., 21 Dec 2025).
3. Model Architecture and Algorithmic Design
PIPCFR is implemented as an end-to-end trainable architecture comprising three main modules:
- Post-treatment Representation Learner (PRL) (): An MLP mapping to .
- Pseudo Counterfactual Constructor (PCC) (): A TARNet-style network producing , serving as a “teacher.”
- Counterfactual Regressor (CFR) (): Another TARNet-style network as the “student” for ITE estimation.
The training objective decomposes into:
- Propensity/independence losses: (propensity model log-likelihood) and (KL-divergence penalizing dependence between and ).
- Teacher loss: , balancing fit to outcomes with IPM/MMD regularization to match representation distributions across treatments.
- PIP student loss: incorporates fit to factual outcome and imputed counterfactual, with a cross-term to reduce estimation variance.
The model is trained using alternating minimization:
- Update with ;
- Update with ;
- Update with ;
- Update with .
Post-treatment features are only needed at training; at test time, only is used for prediction (Lin et al., 21 Dec 2025).
4. Empirical Results and Benchmarking
PIPCFR was evaluated on a suite of standard datasets:
- IHDP (RCT-based simulation, 747 units),
- News (5000 items, high-dimensional word counts with topic model post-treatment embeddings),
- Large synthetic temporal causal system (10,000 units, mediator/adjustment/noise components),
- Real-world gaming data (3M users, high-dimensional input).
Comparisons were made against Causal Forest, meta-learners (XLearner, RLearner, DR-Learner), representation methods (TARNet, CFRNet-MMD/Wass, DRCFR, ESCFR), DragonNet, and matching-based techniques.
Performance is reported using root PEHE on held-out data. PIPCFR_WASS achieves
- $2.35$ vs. baseline (↓20%) on IHDP,
- $0.44$ vs. (↓42%) on News,
- $3.06$ vs. (↓25%) on synthetic data,
- points reduction vs. best baseline on real-world gaming data.
PIPCFR demonstrates improved robustness as post-treatment noise grows and as the length of post-treatment sequences increases, with advantages increasing with the difficulty of post-treatment structure. Substituting alternative teacher networks (DRCFR, DragonNet, ESCFR) preserves relative performance gains (Lin et al., 21 Dec 2025).
5. Connections to Pseudo-outcome Counterfactual Regression
The idea of learning from pseudo-outcomes for counterfactual inference predates PIPCFR and has been formalized in the Forster–Warmuth framework (Yang et al., 2023). The generic strategy involves constructing a pseudo-outcome via efficient influence functions so that its conditional mean yields the target counterfactual regression, and then fitting a flexible regressor (e.g., series estimator). A central insight from this literature is that, with appropriate construction, pseudo-outcome plug-in estimators can achieve second-order bias in the error of nuisance estimation and minimax rate optimality under weak regularity conditions. When post-treatment variables are available, they can be incorporated in the pseudo-outcome and in the nuisance estimation stage, provided suitable identifiability (e.g., sequential ignorability) holds (Yang et al., 2023). PIPCFR elaborates this principle by learning data-driven representations of post-treatment variables, providing explicit variance/bias trade-off, and connecting the integration of post-treatment information to ITE estimation risk via a formal bound.
6. Limitations and Future Directions
PIPCFR is subject to several constraints. It currently assumes binary treatment assignment, complete post-treatment data available during training, and relatively simple MLP-based sequence models for post-treatment representation. The framework relies fundamentally on no hidden confounding and overlap. Directions for future research include:
- Generalization to multi-armed or continuous treatments,
- Identification in the presence of hidden confounders (e.g., via proxy-variable or instrumental-variable methods),
- Deployment of advanced sequence models (RNNs, Transformers) for richer post-treatment data,
- Theoretical refinement of bias–variance trade-offs in the learned post-treatment representation (Lin et al., 21 Dec 2025).
These avenues are critical to extending the practical scope and theoretical guarantees of pseudo-outcome imputation with post-treatment variables.
7. Summary Table: Core Components of PIPCFR
| Component | Function | Technical Detail |
|---|---|---|
| Post-treatment Representation Learner | Learns from | MLP, enforced independence with KL/IPM regularizer |
| Pseudo Counterfactual Constructor (PCC) | Imputes pseudo-counterfactual | TARNet-style “teacher” network |
| Counterfactual Regressor (CFR) | Final ITE predictor | TARNet-style “student” network |
The design enables PIPCFR to systematically exploit outcome-relevant post-treatment information while rigorously controlling induced biases to yield improved counterfactual estimation fidelity (Lin et al., 21 Dec 2025).