Towards Stable Machine Learning Model Retraining via Slowly Varying Sequences (2403.19871v5)
Abstract: We consider the problem of retraining ML models when new batches of data become available. Existing approaches greedily optimize for predictive power independently at each batch, without considering the stability of the model's structure or analytical insights across retraining iterations. We propose a model-agnostic framework for finding sequences of models that are stable across retraining iterations. We develop a mixed-integer optimization formulation that is guaranteed to recover Pareto optimal models (in terms of the predictive power-stability trade-off) with good generalization properties, as well as an efficient polynomial-time algorithm that performs well in practice. We focus on retaining consistent analytical insights-which is important to model interpretability, ease of implementation, and fostering trust with users-by using custom-defined distance metrics that can be directly incorporated into the optimization problem. We evaluate our framework across models (regression, decision trees, boosted trees, and neural networks) and application domains (healthcare, vision, and language), including deployment in a production pipeline at a major US hospital. We find that, on average, a 2% reduction in predictive power leads to a 30% improvement in stability.
- Towards a rigorous science of interpretable machine learning. arXiv: Machine Learning, 2017.
- Definitions, methods, and applications in interpretable machine learning. Proceedings of the National Academy of Sciences, 116(44):22071–22080, 2019.
- Zachary C. Lipton. The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue, 16(3):31–57, 2018.
- Algorithm, human, or the centaur: How to enhance clinical care? SSRN Electronic Journal, 2022.
- Matthew Hindman. Building better models: Prediction, replication, and machine learning in the social sciences. The ANNALS of the American Academy of Political and Social Science, 659(1):48–62, 2015.
- Interpretable machine learning in healthcare. In Proceedings of the 2018 ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, BCB ’18, pages 559–560, New York, NY, USA, 2018. Association for Computing Machinery.
- Decision tree instability and active learning. In Joost N. Kok, Jacek Koronacki, Raomon Lopez de Mantaras, Stan Matwin, Dunja Mladenič, and Andrzej Skowron, editors, Machine Learning: ECML 2007, pages 128–139, Berlin, Heidelberg, 2007. Springer Berlin Heidelberg.
- Continual learning in practice. In NeurIPS 2018, 2018.
- Streammlops: Operationalizing online learning for big data streaming & real-time applications. pages 3508–3521, 2023.
- Matchmaker: Data drift mitigation in machine learning for large-scale systems. In D. Marculescu, Y. Chi, and C. Wu, editors, Proceedings of Machine Learning and Systems, volume 4, pages 77–94, 2022.
- Adaptation strategies for automated machine learning on evolving data. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(9):3067–3078, 2021.
- Repairing without retraining: Avoiding disparate impact with counterfactual distributions. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 6618–6627. PMLR, 2019.
- DeltaGrad: Rapid retraining of machine learning models. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 10355–10366. PMLR, 2020.
- Ekaba Bisong. Batch vs. Online Learning, pages 199–201. Apress, Berkeley, CA, 2019.
- Slowly varying regression under sparsity. arXiv preprint arXiv:2102.10773, 2021.
- Improving stability in decision tree models, 2023.
- Iterative random forests to discover predictive and stable high-order interactions. Proceedings of the National Academy of Sciences, 115(8):1943–1948, 2018.
- A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, pages 4768–4777, Red Hook, NY, USA, 2017. Curran Associates Inc.
- Patient outcome predictions improve operations at a large hospital network, 2023.