DeepROCK: Error-controlled interaction detection in deep neural networks (2309.15319v1)
Abstract: The complexity of deep neural networks (DNNs) makes them powerful but also makes them challenging to interpret, hindering their applicability in error-intolerant domains. Existing methods attempt to reason about the internal mechanism of DNNs by identifying feature interactions that influence prediction outcomes. However, such methods typically lack a systematic strategy to prioritize interactions while controlling confidence levels, making them difficult to apply in practice for scientific discovery and hypothesis validation. In this paper, we introduce a method, called DeepROCK, to address this limitation by using knockoffs, which are dummy variables that are designed to mimic the dependence structure of a given set of features while being conditionally independent of the response. Together with a novel DNN architecture involving a pairwise-coupling layer, DeepROCK jointly controls the false discovery rate (FDR) and maximizes statistical power. In addition, we identify a challenge in correctly controlling FDR using off-the-shelf feature interaction importance measures. DeepROCK overcomes this challenge by proposing a calibration procedure applied to existing interaction importance measures to make the FDR under control at a target level. Finally, we validate the effectiveness of DeepROCK through extensive experiments on simulated and real datasets.
- Z. Obermeyer and E. J. Emanuel. Predicting the future–big data, machine learning, and clinical medicine. The New England Journal of Medicine, 375(13):1216, 2016.
- Z. C. Lipton. The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue, 16(3):31–57, 2018.
- Explaining deep neural networks and beyond: A review of methods and applications. Proceedings of the IEEE, 109(3):247–278, 2021.
- Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013.
- Learning important features through propagating activation differences. In International Conference on Machine Learning, 2017.
- A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems, 2017.
- Axiomatic attribution for deep networks. In International Conference on Machine Learning, 2017.
- DANCE: Enhancing saliency maps using decoys. In International Conference on Machine Learning, 2021.
- Detecting statistical interactions from neural network weights. International Conference on Learning Representations, 2018.
- A wider field of view to predict expression. Nature Methods, 18(10):1155–1156, 2021.
- D. S. Watson. Interpretable machine learning for genomics. Human Genetics, 141(9):1499–1513, 2022.
- Interpretable artificial intelligence through the lens of feature interaction. arXiv preprint arXiv:2103.03103, 2021.
- Interpretation of neural networks is fragile. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 3681–3688, 2019.
- The (un)reliability of saliency methods. Explainable AI: Interpreting, explaining and visualizing deep learning, pages 267–280, 2019.
- ACE: Explaining cluster from an adversarial perspective. In International Conference on Machine Learning, 2021.
- Y. Benjamini and Y. Hochberg. Controlling the false discovery rate: a practical and powerful approach to multiple testing. Journal of the Royal Statistical Society Series B, 57:289–300, 1995.
- Controlling the false discovery rate via knockoffs. The Annals of Statistics, 43(5):2055–2085, 2015.
- Panning for gold: Model-X knockoffs for high-dimensional controlled variable selection. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 80(3):551–577, 2018.
- DeepPINK: reproducible feature selection in deep neural networks. In Advances in Neural Information Processing Systems, 2018.
- Detecting statistical interactions with additive groves of trees. In International Conference on Machine learning, pages 1000–1007, 2008.
- Y. Benjamini and D. Yekutieli. The control of the false discovery rate in multiple testing under dependency. The Annals of Statistics, 29:1165–1188, 2001.
- Strong control, conservative point estimation, and simultaneous conservative consistency of false discovery rates: A unified approach. Journal of the Royal Statistical Society, Series B, 66:187–205, 2004.
- Adapting to unknown sparsity by controlling the false discovery rate. The Annals of Statistics, 34:584–653, 2006.
- To how many simultaneous hypothesis tests can normal, student’s t𝑡titalic_t or bootstrap calibration be applied? Journal of the American Statistical Association, 102:1282–1288, 2007.
- W. B. Wu. On false discovery control under dependence. The Annals of Statistics, 36:364–380, 2008.
- S. Clarke and P. Hall. Robustness of multiple testing procedures against dependence. The Annals of Statistics, 37:332–358, 2009.
- P. Hall and Q. Wang. Strong approximations of level exceedences related to multiple hypothesis testing. Bernoulli, 16:418–434, 2010.
- Control of the false discovery rate under arbitrary covariance dependence (with discussion). Journal of the American Statistical Association, 107:1019–1045, 2012.
- KnockoffGAN: Generating knockoffs for feature selection using generative adversarial networks. In International Conference on Learning Representations, 2018.
- Y. Liu and C. Zheng. Auto-encoding knockoff generator for FDR controlled variable selection. arXiv preprint arXiv:1809.10765, 2018.
- Deep knockoffs. Journal of the American Statistical Association, 115(532):1861–1872, 2020.
- Deep direct likelihood knockoffs. In Advances in Neural Information Processing Systems, volume 33, pages 5036–5046, 2020.
- Neural interaction transparency (NIT): Disentangling learned interactions for improved interpretability. Advances in Neural Information Processing Systems, 31, 2018.
- GAMI-Net: An explainable neural network based on generalized additive models with structured interactions. Pattern Recognition, 120:108192, 2021.
- Learning global pairwise interactions with bayesian neural networks. European Conference on Artificial Intelligence, 2020.
- Towards interaction detection using topological analysis on neural networks. Advances in Neural Information Processing Systems, 2020.
- A. Badre and C. Pan. LINA: A linearizing neural network architecture for accurate first-order and second-order interpretations. IEEE Access, 10:36166–36176, 2022.
- G. D. Garson. Interpreting neural-network connection weights. AI Expert, 6(4):46–51, 1991.
- Recovering pairwise interactions using neural networks. arXiv preprint arXiv:1901.08361, 2019.
- From local explanations to global understanding with explainable AI for trees. Nature Machine Intelligence, 2(1):56–67, 2020.
- How does this interaction affect me? interpretable attribution for feature interactions. Advances in Neural Information Processing Systems, 33:6147–6159, 2020.
- Explaining explanations: Axiomatic feature interactions for deep networks. Journal of Machine Learning Research, 22:104:1–104:54, 2021.
- The shapley taylor interaction index. In International Conference on Machine Learning, pages 9259–9268. PMLR, 2020.
- NODE-GAM: Neural generalized additive model for interpretable deep learning. International Conference on Learning Representations, 2022.
- Explaining local, global, and higher-order interactions in deep learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1224–1233, 2021.
- Interpreting multivariate shapley interactions in DNNs. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 10877–10886, 2021.
- Improving performance of deep learning models with axiomatic attribution priors and expected gradients. Nature Machine Intelligence, 3(7):620–631, 2021.
- False discovery rate estimation for cross-linked peptides identified by mass spectrometry. Nature Methods, 9(9):901–903, 2012.
- Iterative random forests to discover predictive and stable high-order interactions. Proceedings of the National Academy of Sciences, 115(8):1943–1948, 2018.
- Transcription factors bind thousands of active and inactive regions in the Drosophila blastoderm. PLoS Biology, 6(2):e27, 2008.
- TFLink: an integrated gateway to access transcription factor–target gene interactions for multiple species. Database, 2022:baac083, 2022.
- Plan and operation of the NHANES I Epidemiologic Followup Study, 1992. 1997.
- The blood urea nitrogen/creatinine (BUN/cre) ratio was U-shaped associated with all-cause mortality in general population. Renal Failure, 44(1):184–190, 2022.
- Association of serum potassium with all-cause mortality in patients with and without heart failure, chronic kidney disease, and/or diabetes. American Journal of Nephrology, 46(3):213–221, 2017.
- Blood urea nitrogen is independently associated with renal outcomes in Japanese patients with stage 3–5 chronic kidney disease: a prospective observational study. BMC Nephrology, 20:1–10, 2019.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.