Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

When does Subagging Work? (2404.01832v1)

Published 2 Apr 2024 in stat.ML and cs.LG

Abstract: We study the effectiveness of subagging, or subsample aggregating, on regression trees, a popular non-parametric method in machine learning. First, we give sufficient conditions for pointwise consistency of trees. We formalize that (i) the bias depends on the diameter of cells, hence trees with few splits tend to be biased, and (ii) the variance depends on the number of observations in cells, hence trees with many splits tend to have large variance. While these statements for bias and variance are known to hold globally in the covariate space, we show that, under some constraints, they are also true locally. Second, we compare the performance of subagging to that of trees across different numbers of splits. We find that (1) for any given number of splits, subagging improves upon a single tree, and (2) this improvement is larger for many splits than it is for few splits. However, (3) a single tree grown at optimal size can outperform subagging if the size of its individual trees is not optimally chosen. This last result goes against common practice of growing large randomized trees to eliminate bias and then averaging to reduce variance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. Recursive partitioning for heterogeneous causal effects. Proceedings of the National Academy of Sciences, 113(27):7353–7360, 2016.
  2. Leo Breiman. Bagging predictors. Machine Learning, 24:123–140, 1996.
  3. Leo Breiman. Random forests. Machine Learning, 45:5–32, 2001.
  4. Classification and regression trees. CRC Press, 1984.
  5. Analyzing bagging. Annals of Statistics, 30(4):927–961, 2002.
  6. Observations on bagging. Statistica Sinica, 16(2):323–351, 2006.
  7. Why do random forests work? understanding tree ensembles as self-regularizing adaptive smoothers. doi:10.48550/arXiv.2402.01502, 2024.
  8. Machine learning bias, statistical bias, and statistical variance of decision tree algorithms. Working Paper, 1995.
  9. Impact of subsampling and tree depth on random forests. ESAIM: Probability and Statistics, 22:96–128, 2018.
  10. On bagging and nonlinear estimation. Journal of statistical planning and inference, 137(3):669–683, 2007.
  11. Yves Grandvalet. Bagging equalizes influence. Machine Learning, 55:251–270, 2004.
  12. Yves Grandvalet. Stability of bagged decision trees. In Proceedings of the XLIII Scientific Meeting of the Italian Statistical Society, pages 221–230. CLEUP, 2006.
  13. A distribution-free theory of nonparametric regression, volume 1. Springer, 2002.
  14. The elements of statistical learning: data mining, inference, and prediction, volume 2. Springer, 2009.
  15. Wassily Hoeffding. A class of statistics with asymptotically normal distribution. Breakthroughs in Statistics: Foundations and Basic Theory, pages 308–334, 1992.
  16. An introduction to statistical learning, volume 112. Springer, 2013.
  17. Large scale prediction with decision trees. Journal of the American Statistical Association, 119(545):1–27, 2023.
  18. Classification and regression by randomforest. R News, 2(3):18–22, 2002.
  19. Yi Lin and Yongho Jeon. Random forests and adaptive nearest neighbors. Journal of the American Statistical Association, 101(474):578–590, 2006.
  20. Quantifying uncertainty in random forests via confidence intervals and hypothesis tests. Journal of Machine Learning Research, 17(26):1–41, 2016.
  21. Elizbar A Nadaraya. On estimating regression. Theory of Probability & Its Applications, 9(1):141–142, 1964.
  22. Consistency of random forests. Annals of Statistics, 43(4):1716 – 1741, 2015.
  23. Mark R Segal. Machine learning benchmarks and random forest regression. Working Paper, 2004.
  24. Bagging provides assumption-free stability. doi:10.48550/arXiv.2301.12600, 2023.
  25. Charles J Stone. Consistent nonparametric regression. Annals of Statistics, 5(4):595–620, 1977.
  26. Aad W Van der Vaart. Asymptotic statistics, volume 3. Cambridge University Press, 2000.
  27. Estimation and inference of heterogeneous treatment effects using random forests. Journal of the American Statistical Association, 113(523):1228–1242, 2018.
  28. Geoffrey S Watson. Smooth regression analysis. Sankhyā: Indian Journal of Statistics, Series A, 26(4):359–372, 1964.
  29. Trees, forests, chickens, and eggs: when and why to prune trees in a random forest. Statistical Analysis and Data Mining: The ASA Data Science Journal, 16(1):45–64, 2023.

Summary

We haven't generated a summary for this paper yet.