Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
143 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A joint test of unconfoundedness and common trends (2404.16961v3)

Published 25 Apr 2024 in econ.EM

Abstract: This paper introduces an overidentification test of two alternative assumptions to identify the average treatment effect on the treated in a two-period panel data setting: unconfoundedness and common trends. Under the unconfoundedness assumption, treatment assignment and post-treatment outcomes are independent, conditional on control variables and pre-treatment outcomes, which motivates including pre-treatment outcomes in the set of controls. Conversely, under the common trends assumption, the trend and the treatment assignment are independent, conditional on control variables. This motivates employing a Difference-in-Differences (DiD) approach by comparing the differences between pre- and post-treatment outcomes of the treatment and control group. Given the non-nested nature of these assumptions and their often ambiguous plausibility in empirical settings, we propose a joint test using a doubly robust statistic that can be combined with machine learning to control for observed confounders in a data-driven manner. We discuss various causal models that imply the satisfaction of either common trends, unconfoundedness, or both assumptions jointly, and we investigate the finite sample properties of our test through a simulation study. Additionally, we apply the proposed method to five empirical examples using publicly available datasets and find the test to reject the null hypothesis in two cases.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. Semiparametric difference-in-differences estimators. Review of Economic Studies 72, 1–19.
  2. Estimating the effect of training programms on earnings. The Review of Economics and Statistics 6, 47–57.
  3. causalsens: Selection Bias Approach to Sensitivity Analysis for Causal Effects. URL: https://CRAN.R-project.org/package=causalsens. r package version 0.1.2.
  4. Should we combine difference in differences with conditioning on pre-treatment outcomes. working paper, Toulouse School of Economics .
  5. Double/debiased machine learning for difference-in-differences models. The Econometrics Journal 23, 177–191.
  6. Double/debiased machine learning for treatment and structural parameters. The Econometrics Journal 21, C1–C68.
  7. Planning of Experiments. Wiley, New York.
  8. Causal effects in non-experimental studies: Reevaluating the evaluation of training programmes. Journal of American Statistical Association 94, 1053–1062.
  9. Propensity-score-matching methods for nonexperimental causal studies. The Review of Economics and Statistics 84, 151–161.
  10. Selection and parallel trends. arXiv preprint 2203.09001 .
  11. The common structure of statistical models of truncation, sample selection and limited dependent variables and a simple estimator for such models. Annals of Economic and Social Measurement 5, 475–492.
  12. A generalization of sampling without replacement from a finite universe. Journal of the American Statistical Association 47, 663–685.
  13. Testing the identification of causal effects in data. arXiv preprint 2203.15890 .
  14. Nonparametric estimation of average treatment effects under exogeneity: a review. The Review of Economics and Statistics 86, 4–29.
  15. Double machine learning based program evaluation under unconfoundedness arXiv:arXiv preprint 2003.03191.
  16. Algorithm-agnostic significance testing in supervised learning with multimodal data. arXiv preprint 2402.14416 .
  17. Evaluating the econometric evaluations of training programs with experimental data. American Economic Review 76, 604–620.
  18. The estimation of causal effects by difference-in-difference methods. Foundations and Trends in Econometrics 4, 165–224.
  19. On the application of probability theory to agricultural experiments. essay on principles. Statistical Science Reprint, 5, 463–480.
  20. Probabilistic reasoning in intelligent systems: networks of plausible inference. Morgan Kaufmann, San Mateo.
  21. Causality: Models, Reasoning, and Inference. Cambridge University Press, Cambridge.
  22. Consistent significance testing for nonparametric regression. Journal of Business & Economic Statistics 15, 369–378.
  23. Semiparametric efficiency in multivariate regression models with missing data. Journal of the American Statistical Association 90, 122–129.
  24. Estimation of regression coefficients when some regressors are not always observed. Journal of the American Statistical Association 90, 846–866.
  25. Analysis of semiparametric regression models for repeated outcomes in the presence of missing data. Journal of the American Statistical Association 90, 106–121.
  26. The central role of the propensity score in observational studies for causal effects. Biometrika 70, 41–55.
  27. Constructing a control group using multivariate matched sampling methods that incorporate the propensity score. The American Statistician 39, 33–38.
  28. When is parallel trends sensitive to functional form? Econometrica 91, 737–747.
  29. What’s trending in difference-in-differences? a synthesis of the recent econometrics literature. arXiv preprint 2201.01194 .
  30. Comment on ’randomization analysis of experimental data: The fisher randomization test’ by d. basu. Journal of American Statistical Association 75, 591–593.
  31. Estimating causal effects of treatments in randomized and nonrandomized studies. Journal of Educational Psychology 66, 688–701.
  32. Does matching overcome lalonde’s critique of nonexperimental estimators? Journal of Econometrics 125, 305–353.
  33. On the Mode of Communication of Cholera.
  34. Causation, prediction, and search. Springer.
  35. Regresson shrinkage and selection via the lasso. Journal of the Royal Statistical Society 58, 267–288.
  36. Econometric Analysis of Cross Section and Panel Data. MIT Press, Cambridge.
  37. Efficient difference-in-differences estimation with high-dimensional common trend confounding. arXiv preprint 1809.01643 .

Summary

  • The paper introduces a novel joint test for the crucial causal inference assumptions of unconfoundedness and common trends in two-period panel data analysis.
  • It proposes a flexible doubly robust statistic augmented with machine learning to estimate confounder effects and provide robustness against model misspecification.
  • Empirical validation shows the test's effectiveness in identifying violations of these assumptions, offering researchers a robust tool for validating causal inferences.

Overview of a Joint Test for Unconfoundedness and Common Trends

This paper introduces a methodological framework for the overidentification test of two fundamental assumptions—unconfoundedness and common trends—to identify the average treatment effect on the treated (ATET) in a two-period panel data setting. The discussion delineates the assumptions underlying two prevalent techniques in treatment effect estimation: inclusion of pre-treatment outcomes as controls and the Difference-in-Differences (DiD) approach. It also addresses the conceptual non-nested nature of these assumptions, presenting a novel joint test leveraging a doubly robust statistic, augmented with machine learning for enhanced data-driven control of confounders.

Key Conceptual Foundations

  1. Unconfoundedness: The unconfoundedness assumption posits that conditional on control variables and pre-treatment outcomes, the treatment assignment is independent of post-treatment outcomes. This assumption underpins using pre-treatment outcomes as controls in causal inference.
  2. Common Trends: The common trends assumption underlies the DiD approach. It assumes that, conditioned on covariates, treatment assignment is independent of the time trends in outcomes, which facilitates leveraging changes in pre- and post-treatment outcomes among treatment and control groups to identify the ATET.

In empirical research, determining which assumption is more plausible is not always clear, motivating the need for a robust testing methodology. This paper fills this gap by proposing a doubly robust testing procedure.

Methodological Contributions

The central contribution of the paper is the development of a joint test for the unconfoundedness and common trends assumptions using a flexible doubly robust statistic that can be coupled with machine learning to estimate confounder effects. The proposed doubly robust estimator provides consistent estimates of ATET when either the outcome models or the propensity scores are correctly specified, providing robustness against misspecifications.

  • Doubly Robust Statistics: The paper emphasizes the orthogonality properties of the doubly robust statistic, ensuring asymptotic normality and consistency for large samples even when nuisance parameters are machine-learned. This facilitates the robust estimation of test statistics in high-dimensional settings.
  • Implementation with Machine Learning (ML): The methodology leverages ML-based estimators for conditional expectation functions and propensity scores, increasing the adaptability and efficacy of controlling confounding variables.

Empirical Assessment

The paper evaluates the finite-sample properties of the proposed test through simulation studies and applies the method to real-world datasets, including publicly available datasets with different experimental designs and assumptions about treatment assignments. These practical applications validate the effectiveness of the method, demonstrating rejection of the null hypothesis in specific cases, signifying violations of either or both assumptions.

Implications and Future Directions

The results hold significant implications for empirical research, suggesting a more nuanced and robust means of validating treatment effect estimation when underlying assumptions are ambiguous. The methodological framework presented could guide empirical analysts in a variety of fields—from economics to epidemiology—in strengthening causal inferences drawn from panel data.

Future research directions proposed include extending this framework to multi-period data, facilitating more comprehensive analyses of policy interventions and natural experiments. Additionally, further integration with advanced machine learning methods could enhance the flexibility and scalability of the approach in processing increasingly complex datasets.

In summary, this paper provides a rigorous and adaptable framework for improving causal inference in panel data, advocating for empirically grounded validation of foundational assumptions through the innovative application of joint testing and machine learning techniques.