Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Smooth Min-Max Monotonic Networks (2306.01147v3)

Published 1 Jun 2023 in cs.LG and cs.AI

Abstract: Monotonicity constraints are powerful regularizers in statistical modelling. They can support fairness in computer-aided decision making and increase plausibility in data-driven scientific models. The seminal min-max (MM) neural network architecture ensures monotonicity, but often gets stuck in undesired local optima during training because of partial derivatives of the MM nonlinearities being zero. We propose a simple modification of the MM network using strictly-increasing smooth minimum and maximum functions that alleviates this problem. The resulting smooth min-max (SMM) network module inherits the asymptotic approximation properties from the MM architecture. It can be used within larger deep learning systems trained end-to-end. The SMM module is conceptually simple and computationally less demanding than state-of-the-art neural networks for monotonic modelling. Our experiments show that this does not come with a loss in generalization performance compared to alternative neural and non-neural approaches.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. M. J. Best and N. Chakravarti. Active set algorithms for isotonic regression; a unifying framework. Mathematical Programming, 47(1-3):425–439, 1990.
  2. Monotonic classification: An overview on algorithms, performance measures and data sets. Neurocomputing, 341:168–182, 2019.
  3. A similarity-based QSAR model for predicting acute toxicity towards the fathead minnow (pimephales promelas). SAR and QSAR in Environmental Research, 26(3):217–243, 2015.
  4. T. Chen and C. Guestrin. XGBoost: A scalable tree boosting system. In International Conference on Knowledge Discovery and Data Mining (KDD), pages 785–794. ACM, 2016.
  5. Fast and accurate deep network learning by exponential linear units (ELUs). In International Conference on Learning Representations (ICLR), 2016.
  6. Avoiding resentment via monotonic fairness. arXiv preprint arXiv:1909.01251, 2019.
  7. H. Daniels and M. Velikova. Monotone and partially monotone neural networks. IEEE Transactions on Neural Networks, 21(6):906–917, 2010.
  8. Isotone optimization in R: pool-adjacent-violators algorithm (PAVA) and active set methods. Journal of Statistical Software, 32(5):1–24, 2009.
  9. D. Dua and C. Graff. UCI machine learning repository, 2017. URL http://archive.ics.uci.edu/ml.
  10. How to incorporate monotonicity in deep networks while preserving flexibility? In NeurIPS 2019 Workshop on Machine Learning with Guarantees, 2019.
  11. Allometric equations to estimate the dry mass of sahel woody plants from very-high resolution satellite imagery. Forest Ecology and Management, 529, 2023.
  12. Certified monotonic neural networks. Advances in Neural Information Processing Systems (NeurIPS), 33:15427–15438, 2020.
  13. Rectifier nonlinearities improve neural network acoustic models. In International Conference on Machine Learning (ICML), 2013.
  14. D. Mikulincer and D. Reichman. Size and depth of monotone neural networks: interpolation and approximation. In Advances in Neural Information Processing Systems (NeurIPS), 2022.
  15. Fast and flexible monotonic functions with ensembles of lattices. In Advances in Neural Information Processing Systems (NeurIPS), volume 29, 2016.
  16. V. Nair and G. E. Hinton. Rectified linear units improve restricted Boltzmann machines. In International Conference on Machine Learning (ICML), pages 807–814, 2010.
  17. A. Niculescu-Mizil and R. Caruana. Predicting good probabilities with supervised learning. In International Conference on Machine learning (ICML), pages 625–632, 2005.
  18. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
  19. L. Prechelt. Early stopping — but when? In G. Montavon, G. B. Orr, and K.-R. Müller, editors, Neural Networks: Tricks of the Trade: Second Edition, pages 53–67. Springer, 2012. ISBN 978-3-642-35289-8.
  20. M. Riedmiller and H. Braun. A direct adaptive method for faster backpropagation learning: The RPROP algorithm. In IEEE International Conference on Neural Networks, pages 586–591. IEEE, 1993.
  21. J. Sill. Monotonic networks. In M. Jordan, M. Kearns, and S. Solla, editors, Advances in Neural Information Processing Systems (NeurIPS), volume 10. MIT Press, 1997.
  22. Counterexample-guided learning of monotonic neural networks. Advances in Neural Information Processing Systems (NeurIPS), 33:11936–11948, 2020.
  23. A. Tsanas and A. Xifara. Accurate quantitative estimation of energy performance of residential buildings using statistical machine learning tools. Energy and Buildings, 49:560–567, 2012.
  24. Sub-continental scale carbon stocks of individual trees in African drylands. Nature, 615:80–86, 2023.
  25. S. Wang and M. Gupta. Deontological ethics by monotonicity shape constraints. In International Conference on Artificial Intelligence and Statistics (AISTATS), pages 2043–2054, 2020.
  26. Hierarchical lattice layer for partially monotone neural networks. In Advances in Neural Information Processing Systems (NeurIPS), 2022.
  27. I.-C. Yeh. Modeling of strength of high-performance concrete using artificial neural networks. Cement and Concrete Research, 28(12):1797–1808, 1998.
  28. Deep lattice networks and partial monotonic functions. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems (NeurIPS), volume 30, 2017.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com