Full Bayesian Significance Testing for Neural Networks (2401.13335v1)
Abstract: Significance testing aims to determine whether a proposition about the population distribution is the truth or not given observations. However, traditional significance testing often needs to derive the distribution of the testing statistic, failing to deal with complex nonlinear relationships. In this paper, we propose to conduct Full Bayesian Significance Testing for neural networks, called \textit{n}FBST, to overcome the limitation in relationship characterization of traditional approaches. A Bayesian neural network is utilized to fit the nonlinear and multi-dimensional relationships with small errors and avoid hard theoretical derivation by computing the evidence value. Besides, \textit{n}FBST can test not only global significance but also local and instance-wise significance, which previous testing methods don't focus on. Moreover, \textit{n}FBST is a general framework that can be extended based on the measures selected, such as Grad-\textit{n}FBST, LRP-\textit{n}FBST, DeepLIFT-\textit{n}FBST, LIME-\textit{n}FBST. A range of experiments on both simulated and real data are conducted to show the advantages of our method.
- Layer-wise relevance propagation for neural networks with local renormalization layers. In International Conference on Artificial Neural Networks, 63β71. Springer.
- Variational inference: A review for statisticians. Journal of the American statistical Association, 112(518): 859β877.
- Weight uncertainty in neural networks. arXiv preprint arXiv:1505.05424.
- AlphaPortfolio: Direct construction through deep reinforcement learning and interpretable AI. Available at SSRN 3554486.
- Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems. In 2016 IEEE symposium on security and privacy (SP), 598β617. IEEE.
- Can a significance test be genuinely Bayesian? Bayesian Analysis, 3(1): 79 β 100.
- Evidence and credibility: Full Bayesian significance test for precise hypotheses. Entropy, 1(4): 99β110.
- Efron, B. 1979. Bootstrap Methods: Another Look at the Jackknife. The Annals of Statistics, 7(1): 1 β 26.
- Consistent model specification tests: omitted variables and semiparametric functional forms. Econometrica: Journal of the econometric society, 865β890.
- Fisher, R.Β A. 1922. On the interpretation of Οπ\chiitalic_Ο 2 from contingency tables, and the calculation of P. Journal of the royal statistical society, 85(1): 87β94.
- Friedman, J.Β H. 2001. Greedy function approximation: a gradient boosting machine. Annals of statistics, 1189β1232.
- Gozalo, P.Β L. 1993. A consistent model specification test for nonparametric estimation of regression function models. Econometric Theory, 9(3): 451β477.
- A simple and effective model-based variable importance measure. arXiv preprint arXiv:1805.04755.
- Learning stable graphs from multiple environments with selection bias. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2194β2202.
- Significance tests for neural networks. Journal of Machine Learning Research, 21(227): 1β29.
- Multilayer feedforward networks are universal approximators. Neural networks, 2(5): 359β366.
- Bayesian parameter estimation via variational methods. Statistics and Computing, 10(1): 25β37.
- Jeffreys, H. 1998. The theory of probability. OuP Oxford.
- STDEN: Towards physics-guided neural networks for traffic flow prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, volumeΒ 36, 4048β4056.
- Interpretable spatiotemporal deep learning model for traffic flow prediction based on potential energy fields. In 2020 IEEE International Conference on Data Mining (ICDM), 1076β1081. IEEE.
- Bayes factors. Journal of the american statistical association, 90(430): 773β795.
- What uncertainties do we need in bayesian deep learning for computer vision? Advances in neural information processing systems, 30.
- Investigating the influence of noise and distractors on the interpretation of neural networks. arXiv preprint arXiv:1611.07270.
- Nonparametric significance testing. Econometric Theory, 16(4): 576β601.
- Nonparametric selection of regressors: The nonnested case. Econometrica: Journal of the Econometric Society, 207β219.
- A unified approach to interpreting model predictions. Advances in neural information processing systems, 30.
- Analysis of serial measurements in medical research. British Medical Journal, 300(6719): 230β235.
- Illuminating the βblack boxβ: a randomization approach for understanding variable contributions in artificial neural networks. Ecological modelling, 154(1-2): 135β150.
- Orlitzky, M. 2012. How can significance tests be deinstitutionalized? Organizational Research Methods, 15(2): 199β228.
- Bayesian hypothesis testing: an alternative to null hypothesis significance testing (NHST) in psychology and social sciences. In Bayesian inference. IntechOpen.
- Parzen, E. 1962. On estimation of a probability density function and mode. The annals of mathematical statistics, 33(3): 1065β1076.
- Racine, J. 1997. Consistent significance testing for nonparametric regression. Journal of Business & Economic Statistics, 15(3): 369β378.
- β Why should i trust you?β Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, 1135β1144.
- Effect sizes and statistical testing in the determination of clinical significance in behavioral medicine research. Annals of Behavioral Medicine, 27(2): 138β145.
- Scott, D.Β W. 1979. On optimal and data-based histograms. Biometrika, 66(3): 605β610.
- Learning important features through propagating activation differences. In International conference on machine learning, 3145β3153. PMLR.
- Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034.
- Striving for simplicity: The all convolutional net. arXiv preprint arXiv:1412.6806.
- Student. 1908. The probable error of a mean. Biometrika, 6(1): 1β25.
- Axiomatic attribution for deep networks. In International conference on machine learning, 3319β3328. PMLR.
- Accurate quantitative estimation of energy performance of residential buildings using statistical machine learning tools. Energy and buildings, 49: 560β567.
- Vuong, Q.Β H. 1989. Likelihood ratio tests for model selection and non-nested hypotheses. Econometrica: Journal of the Econometric Society, 307β333.
- SVM-based deep stacking networks. In Proceedings of the AAAI conference on artificial intelligence, volumeΒ 33, 5273β5280.
- Traffic flow prediction based on spatiotemporal potential energy fields. IEEE Transactions on Knowledge and Data Engineering.
- Deep fuzzy cognitive maps for interpretable multivariate time series prediction. IEEE transactions on fuzzy systems, 29(9): 2647β2660.
- Multilevel wavelet decomposition network for interpretable time series analysis. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2437β2446.
- Deep Trajectory Recovery with Fine-Grained Calibration using Kalman Filter. IEEE Transactions on Knowledge and Data Engineering, 33(3): 921β934.
- Personalized route recommendation with neural network enhanced search algorithm. IEEE Transactions on Knowledge and Data Engineering, 34(12): 5910β5924.
- Empowering A* search algorithms with neural networks for personalized route recommendation. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, 539β547.
- Interpretability is a kind of safety: An interpreter-based ensemble for adversary defense. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 15β24.
- WHEN: A Wavelet-DTW Hybrid Attention Network for Heterogeneous Time Series Analysis. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2361β2373.
- AlphaStock: A Buying-Winners-and-Selling-Losers Investment Strategy Using Interpretable Deep Reinforcement Attention Networks. In Proceedings of the 25th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD β19, 1900β1908. New York, NY, USA: Association for Computing Machinery. ISBN 9781450362016.
- White, H. 1989a. Learning in artificial neural networks: A statistical perspective. Neural computation, 1(4): 425β464.
- White, H. 1989b. Some asymptotic results for learning in single hidden-layer feedforward network models. Journal of the American Statistical Association, 84(408): 1003β1013.
- Yatchew, A.Β J. 1992. Nonparametric regression tests based on least squares. Econometric Theory, 8(4): 435β451.
- Visualizing and understanding convolutional networks. In European conference on computer vision, 818β833. Springer.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.