Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Don't Waste Your Time: Early Stopping Cross-Validation (2405.03389v2)

Published 6 May 2024 in cs.LG and cs.AI

Abstract: State-of-the-art automated machine learning systems for tabular data often employ cross-validation; ensuring that measured performances generalize to unseen data, or that subsequent ensembling does not overfit. However, using k-fold cross-validation instead of holdout validation drastically increases the computational cost of validating a single configuration. While ensuring better generalization and, by extension, better performance, the additional cost is often prohibitive for effective model selection within a time budget. We aim to make model selection with cross-validation more effective. Therefore, we study early stopping the process of cross-validation during model selection. We investigate the impact of early stopping on random search for two algorithms, MLP and random forest, across 36 classification datasets. We further analyze the impact of the number of folds by considering 3-, 5-, and 10-folds. In addition, we investigate the impact of early stopping with Bayesian optimization instead of random search and also repeated cross-validation. Our exploratory study shows that even a simple-to-understand and easy-to-implement method consistently allows model selection to converge faster; in ~94% of all datasets, on average by ~214%. Moreover, stopping cross-validation enables model selection to explore the search space more exhaustively by considering +167% configurations on average within one hour, while also obtaining better overall performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (41)
  1. Improvement strategies for the f-race algorithm: Sampling design and iterative refinement. In Bartz-Beielstein, T., Blesa Aguilera, M. J., Blum, C., Naujoks, B., Roli, A., Rudolph, G., and Sampels, M., editors, Hybrid Metaheuristics, pages 108–122, Berlin, Heidelberg. Springer Berlin Heidelberg.
  2. AMLTK: A Modular Automl Toolkit in Python. Under Review for Journal of Open Source Software, Issue 6367.
  3. A racing algorithm for configuring metaheuristics. In Proc. of GECCO’02, pages 11–18.
  4. F-race and iterated f-race: An overview. In Bartz-Beielstein, T., Chiarandini, M., Paquete, L., and Preuss, M., editors, Experimental Methods for the Analysis of Optimization Algorithms, pages 311–336, Berlin, Heidelberg. Springer Berlin Heidelberg.
  5. Breiman, L. (2001). Random forests. MLJ, 45:5–32.
  6. Autogluon-tabular: Robust and accurate automl for structured data. arXiv:2003.06505 [stat.ML].
  7. BOHB: Robust and efficient Hyperparameter Optimization at scale. In Proc. of ICML’18, pages 1437–1446.
  8. Auto-Sklearn 2.0: Hands-free automl via meta-learning. JMLR, 23(261):1–61.
  9. Hyperparameter Optimization. In Hutter et al., (2019), chapter 1, pages 3 – 38. Available for free at http://automl.org/book.
  10. Auto-sklearn: Efficient and robust automated machine learning. In Hutter et al., (2019), chapter 6, pages 113–134. Available for free at http://automl.org/book.
  11. OpenML-Python: an extensible Python API for OpenML. JMLR, 22(100):1–5.
  12. Amlb: an automl benchmark. arXiv preprint arXiv:2207.12560.
  13. An open source automl benchmark. In ICML Workshop on AutoML.
  14. Gama: A general automated machine learning assistant. In Dong, Y., Ifrim, G., Mladenić, D., Saunders, C., and Van Hoecke, S., editors, Machine Learning and Knowledge Discovery in Databases. Applied Data Science and Demo Track, pages 560–564, Cham. Springer International Publishing.
  15. Multidimensional scaling by majorization: A review. Journal of Statistical Software, 73:1–26.
  16. Hunter, J. (2007). Matplotlib: A 2d graphics environment. Computing in Science & Engineering, 9(3):90–95.
  17. Sequential model-based optimization for general algorithm configuration. In Proc. of LION’11, pages 507–523.
  18. ParamILS: An automatic algorithm configuration framework. JAIR, 36:267–306.
  19. Automated Machine Learning: Methods, Systems, Challenges. Springer. Available for free at http://automl.org/book.
  20. Non-stochastic best arm identification and Hyperparameter Optimization. In Proc. of AISTATS’16.
  21. Fast cross-validation via sequential testing. J. Mach. Learn. Res., 16(1):1103–1155.
  22. Kruskal, J. B. (1964). Multidimensional scaling by optimizing goodness of fit to a nonmetric hypothesis. Psychometrika, 29(1):1–27.
  23. Hyperband: Bandit-based configuration evaluation for Hyperparameter Optimization. In Proc. of ICLR’17.
  24. Hyperband: A novel bandit-based approach to Hyperparameter Optimization. JMLR, 18(185):1–52.
  25. SMAC3: A versatile bayesian optimization package for Hyperparameter Optimization. JMLR, 23(54):1–9.
  26. The irace package: Iterated racing for automatic algorithm configuration. Operations Research Perspectives, 3:43–58.
  27. The irace package, iterated race for automatic algorithm configuration. Technical report, IRIDIA, Université Libre de Bruxelles, Belgium.
  28. Priorband: Practical hyperparameter optimization in the age of deep learning. Advances in Neural Information Processing Systems, 36.
  29. The racing algorithm: Model selection for lazy learners. Artificial Intelligence Review, 11(1-5):193–225.
  30. McKinney, W. (2010). Data Structures for Statistical Computing in Python. In Stéfan van der Walt and Jarrod Millman, editors, Proceedings of the 9th Python in Science Conference, pages 56–61.
  31. Fast and informative model selection using learning curve cross-validation. IEEE Transactions on Pattern Analysis and Machine Intelligence, PP:1–12.
  32. Scikit-learn: Machine learning in Python. JMLR, 12:2825–2830.
  33. Tabrepo: A large scale repository of tabular model evaluations and its automl applications. arXiv preprint arXiv:2311.02971.
  34. Multi-objective asynchronous successive halving. arXiv:2106.12639 [stat.ML].
  35. Auto-WEKA: combined selection and Hyperparameter Optimization of classification algorithms. In Proc. of KDD’13, pages 847–855.
  36. Lightautoml: Automl solution for a large financial services ecosystem. CoRR, abs/2109.01528.
  37. OpenML: Networked science in machine learning. SIGKDD, 15(2):49–60.
  38. isklearn: Automated machine learning with irace. In 2021 IEEE Congress on Evolutionary Computation (CEC), pages 2354–2361. IEEE.
  39. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nature Methods, 17(3):261–272.
  40. Waskom, M. L. (2021). seaborn: statistical data visualization. Journal of Open Source Software, 6(60):3021.
  41. Quantifying the similarity of algorithm configurations. In Learning and Intelligent Optimization: 10th International Conference, LION 10, Ischia, Italy, May 29–June 1, 2016, Revised Selected Papers 10, pages 203–217. Springer.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com