Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bayesian Additive Regression Networks (2404.04425v1)

Published 5 Apr 2024 in stat.ML and cs.LG

Abstract: We apply Bayesian Additive Regression Tree (BART) principles to training an ensemble of small neural networks for regression tasks. Using Markov Chain Monte Carlo, we sample from the posterior distribution of neural networks that have a single hidden layer. To create an ensemble of these, we apply Gibbs sampling to update each network against the residual target value (i.e. subtracting the effect of the other networks). We demonstrate the effectiveness of this technique on several benchmark regression problems, comparing it to equivalent shallow neural networks, BART, and ordinary least squares. Our Bayesian Additive Regression Networks (BARN) provide more consistent and often more accurate results. On test data benchmarks, BARN averaged between 5 to 20 percent lower root mean square error. This error performance does come at the cost, however, of greater computation time. BARN sometimes takes on the order of a minute where competing methods take a second or less. But, BARN without cross-validated hyperparameter tuning takes about the same amount of computation time as tuned other methods. Yet BARN is still typically more accurate.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. A dynamic ensemble learning algorithm for neural networks. Neural Computing and Applications 32, 8675–8690.
  2. Neural random forests. Sankhya A 81, 347–386.
  3. Bayesian CART model search. Journal of the American Statistical Association 93, 935–948.
  4. BART: Bayesian additive regression trees. The Annals of Applied Statistics 4, 266–298.
  5. Keras. https://keras.io.
  6. UCI machine learning repository. URL: http://archive.ics.uci.edu/ml.
  7. Practical methods of optimization. John Wiley & Sons.
  8. Multivariate adaptive regression splines. The annals of statistics 19, 1–67.
  9. Pricing and hedging derivative securities with neural networks: Bayesian regularization, early stopping, and bagging. IEEE transactions on neural networks 12, 726–734.
  10. Reversible jump markov chain monte carlo computation and bayesian model determination. Biometrika 82, 711–732.
  11. The Elements Of Statistical Learning: Data Mining, Inference, And Prediction. volume 2. Springer.
  12. Genetic algorithm for neural network architecture optimization, in: 2016 3rd International Conference on Logistics Operations Management (GOL), IEEE. pp. 1–4.
  13. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 .
  14. Efficient backprop, in: Orr, G.B., Müller, K.R. (Eds.), Neural networks: tricks of the trade, Springer.
  15. Generalized bayesian additive regression trees models: Beyond conditional conjugacy. arXiv preprint arXiv:2202.09924 .
  16. Bayesian regression tree ensembles that adapt to smoothness and sparsity. Journal of the Royal Statistical Society Series B: Statistical Methodology 80, 1087–1110.
  17. An ensemble of neural networks for weather forecasting. Neural Computing & Applications 13, 112–122.
  18. An ensemble neural network for optimising a cnc milling process. Journal of Manufacturing Systems 71, 377–389.
  19. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems 35, 27730–27744.
  20. Bayesian additive regression trees with model trees. Statistics and Computing 31, 1–13.
  21. Early stopping-but when?, in: Neural Networks: Tricks of the trade. Springer, pp. 55–69.
  22. PyPi Maintainers, 2023. Python package index - PyPi. https://pypi.org/.
  23. Bayclump: Bayesian calibration and temperature reconstructions for clumped isotope thermometry. Authorea Preprints .
  24. Deep learning in neural networks: An overview. Neural networks 61, 85–117.
  25. Grad-cam: Visual explanations from deep networks via gradient-based localization, in: Proceedings of the IEEE international conference on computer vision, pp. 618–626.
  26. Transfer learning-based deep ensemble neural network for plant leaf disease detection. Journal of Plant Diseases and Protection 129, 545–558.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com