Papers
Topics
Authors
Recent
2000 character limit reached

Neural Additive Models for Location Scale and Shape: A Framework for Interpretable Neural Regression Beyond the Mean (2301.11862v2)

Published 27 Jan 2023 in stat.ML and cs.LG

Abstract: Deep neural networks (DNNs) have proven to be highly effective in a variety of tasks, making them the go-to method for problems requiring high-level predictive power. Despite this success, the inner workings of DNNs are often not transparent, making them difficult to interpret or understand. This lack of interpretability has led to increased research on inherently interpretable neural networks in recent years. Models such as Neural Additive Models (NAMs) achieve visual interpretability through the combination of classical statistical methods with DNNs. However, these approaches only concentrate on mean response predictions, leaving out other properties of the response distribution of the underlying data. We propose Neural Additive Models for Location Scale and Shape (NAMLSS), a modelling framework that combines the predictive power of classical deep learning models with the inherent advantages of distributional regression while maintaining the interpretability of additive models. The code is available at the following link: https://github.com/AnFreTh/NAMpy

Definition Search Book Streamline Icon: https://streamlinehq.com
References (54)
  1. Neural additive models: Interpretable machine learning with neural nets. Advances in Neural Information Processing Systems, 34.
  2. Node-gam: Neural generalized additive model for interpretable deep learning. arXiv preprint arXiv:2106.01613.
  3. Generalized gloves of neural additive models: Pursuing transparent and accurate machine learning models in finance. arXiv preprint arXiv:2209.10082.
  4. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, pages 785–794, New York, NY, USA. ACM.
  5. Smoothing reference centile curves: the lms method and penalized likelihood. Statistics in medicine, 11(10):1305–1319.
  6. Generalized additive models from a neural network perspective. In Seventh IEEE International Conference on Data Mining Workshops (ICDMW 2007), pages 265–270. IEEE.
  7. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929.
  8. UCI machine learning repository.
  9. Ngboost: Natural gradient boosting for probabilistic prediction. In International conference on machine learning, pages 2690–2700. PMLR.
  10. Scalable interpretability via polynomials. arXiv preprint arXiv:2205.14108.
  11. Probabilistic Deep Learning: With Python, Keras and TensorFlow Probability. Manning Publications.
  12. Sparse interaction additive networks via feature interaction detection and sparse selection. arXiv preprint arXiv:2209.09326.
  13. FICO (2018). Fico explainable machine learning challenge.
  14. Strictly proper scoring rules, prediction, and estimation. Journal of the American Statistical Association, 102:359–378.
  15. Revisiting deep learning models for tabular data. Advances in Neural Information Processing Systems, 34.
  16. Hastie, T. J. (2017). Generalized additive models. In Statistical models in S, pages 249–307. Routledge.
  17. gamboostlss: An r package for model building and variable selection in the gamlss framework. arXiv preprint arXiv:1407.1774.
  18. Significance tests for neural networks. Journal of Machine Learning Research, 21(227):1–29.
  19. Conditional transformation models. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 76(1):3–27.
  20. Balancing methods for multi-label text classification with long-tailed class distribution. arXiv preprint arXiv:2109.04712.
  21. IBM (2019). Telco customer churn.
  22. Neural additive models for nowcasting. arXiv preprint arXiv:2205.10020.
  23. Kaggle (2019). Churn modelling.
  24. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
  25. Rage against the mean–a review of distributional regression approaches. Econometrics and Statistics.
  26. Lantz, B. (2019). Machine learning with R: expert techniques for predictive modeling. Packt publishing ltd.
  27. Bertgcn: Transductive text classification by combining gcn and bert. arXiv preprint arXiv:2105.05727.
  28. Structural neural additive models: Enhanced interpretable machine learning. arXiv preprint arXiv:2302.09275.
  29. Uncertainty in gradient boosting via ensembles. arXiv preprint arXiv:2006.10562.
  30. Distributional gradient boosting machines. arXiv preprint arXiv:2204.00778.
  31. Attention bottlenecks for multimodal fusion. Advances in Neural Information Processing Systems, 34:14200–14213.
  32. The population biology of abalone (haliotis species) in tasmania. i. blacklip abalone (h. rubra) from the north coast and islands of bass strait. Sea Fisheries Division, Technical Report, 48:p411.
  33. Generalized linear models. Journal of the Royal Statistical Society: Series A (General), 135(3):370–384.
  34. Interpretml: A unified framework for machine learning interpretability. arXiv preprint arXiv:1909.09223.
  35. Sparse spatial autoregressions. Statistics & Probability Letters, 33(3):291–297.
  36. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research, 12:2825–2830.
  37. Extending the neural additive model for survival analysis with ehr data. arXiv preprint arXiv:2211.07814.
  38. Potts, W. J. (1999). Generalized additive neural networks. In Proceedings of the fifth ACM SIGKDD international conference on Knowledge discovery and data mining, pages 194–200.
  39. Neural basis models for interpretability. arXiv preprint arXiv:2205.14120.
  40. ” why should i trust you?” explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1135–1144.
  41. Generalized additive models for location, scale and shape. Journal of the Royal Statistical Society: Series C (Applied Statistics), 54(3):507–554.
  42. Rügamer, D. (2023). A new pho-rmula for improved performance of semi-structured networks. arXiv preprint arXiv:2306.00522.
  43. Semi-structured deep distributional regression: Combining structured additive models and deep learning. arXiv preprint arXiv:2002.05777.
  44. Penalized regression splines in mixture density networks.
  45. Shapley, L. (1953). Quota solutions op n-person games1. Edited by Emil Artin and Marston Morse, page 343.
  46. Modelling rental guide data using mean and dispersion additive models. Journal of the Royal Statistical Society: Series D (The Statistician), 49(4):479–493.
  47. The many shapley values for model explanation. In International conference on machine learning, pages 9269–9278. PMLR.
  48. Mixed poisson regression models with covariate dependent rates. Biometrics, pages 381–400.
  49. Partially interpretable estimators (pie): black-box-refined interpretable machine learning. arXiv preprint arXiv:2105.02410.
  50. Treeflow: Going beyond tree-based gaussian probabilistic regression. arXiv preprint arXiv:2206.04140.
  51. Gami-net: An explainable neural network based on generalized additive models with structured interactions. Pattern Recognition, 120:108192.
  52. Coca: Contrastive captioners are image-text foundation models. arXiv preprint arXiv:2205.01917.
  53. Are transformers effective for time series forecasting? arXiv preprint arXiv:2205.13504.
  54. Film: Frequency improved legendre memory model for long-term time series forecasting. arXiv preprint arXiv:2205.08897.
Citations (9)

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com

GitHub

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 3 likes.

Upgrade to Pro to view all of the tweets about this paper: