An interpretable neural network-based non-proportional odds model for ordinal regression
Abstract: This study proposes an interpretable neural network-based non-proportional odds model (N$3$POM) for ordinal regression. N$3$POM is different from conventional approaches to ordinal regression with non-proportional models in several ways: (1) N$3$POM is defined for both continuous and discrete responses, whereas standard methods typically treat the ordered continuous variables as if they are discrete, (2) instead of estimating response-dependent finite-dimensional coefficients of linear models from discrete responses as is done in conventional approaches, we train a non-linear neural network to serve as a coefficient function. Thanks to the neural network, N$3$POM offers flexibility while preserving the interpretability of conventional ordinal regression. We establish a sufficient condition under which the predicted conditional cumulative probability locally satisfies the monotonicity constraint over a user-specified region in the covariate space. Additionally, we provide a monotonicity-preserving stochastic (MPS) algorithm for effectively training the neural network. We apply N$3$POM to several real-world datasets.
- Agresti, A. (2010). Analysis of Ordinal Categorical Data. John Wiley & Sons.
- Simple ways to interpret effects in modeling ordinal categorical data. Statistica Neerlandica, 72(3):210–223.
- Deep conditional transformation models. In Machine Learning and Knowledge Discovery in Databases. Research Track, pages 3–18. Springer International Publishing.
- Bennett, S. (1983). Log-logistic regression models for survival data. Journal of the Royal Statistical Society. Series C (Applied Statistics), 32(2):165–171.
- An analysis of transformations. Journal of the Royal Statistical Society. Series B (Methodological), 26(2):211–252.
- Inference on counterfactual distributions. Econometrica, 81(6):2205–2268.
- Robust estimation for ordinal regression. Journal of Statistical Planning and Inference, 143(9):1486–1499.
- Cybenko, G. (1989). Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals and Systems, 2(4):303–314.
- Monotone and partially monotone neural networks. IEEE Transactions on Neural Networks, 21(6):906–917.
- UCI Machine Learning Repository. http://archive.ics.uci.edu/ml.
- The conditional distribution of excess returns: An empirical analysis. Journal of the American Statistical Association, 90(430):451–466.
- Deep Learning. MIT Press. http://www.deeplearningbook.org.
- Most likely transformations. Scandinavian Journal of Statistics, 45(1):110–134.
- Estimating conditional distributions with neural networks using R package deeptrafo. arXiv preprint arXiv:2211.13665.
- Deep and interpretable regression models for ordinal outcomes. Pattern Recognition, 122:108263.
- Modeling continuous response variables using ordinal regression. Statistics in Medicine, 36(27):4316–4335.
- Certified monotonic neural networks. In Advances in Neural Information Processing Systems, volume 33, pages 15427–15438.
- Regression models for categorical dependent variables using Stata, volume 7. Stata press.
- Continuously generalized ordinal regression for linear and deep models. In Proceedings of the 2022 SIAM International Conference on Data Mining, pages 28–36.
- McCullagh, P. (1980). Regression models for ordinal data. Journal of Royal Statistical Society. Series B (Methodological), 42(2):109–127.
- Generalized Linear Models. Chapman & Hall CRC, London.
- Partial proportional odds models for ordinal response variables. Journal of Royal Statistical Society. Series C (Applied Statistics), 39(2):205–217.
- Pettitt, A. N. (1984). Proportional odds models for survival data and estimates using ranks. Journal of the Royal Statistical Society: Series C (Applied Statistics), 33(2):169–175.
- Logistic regression model for survival time analysis using time-varying coefficients. American Journal of Mathematical and Management Sciences, 35(4):353–360.
- Sill, J. (1997). Monotonic networks. In Advances in Neural Information Processing Systems, volume 10, pages 661–667.
- Interval censoring and marginal analysis in ordinal regression. Journal of Agricultural Biological and Environmental Statistics, 4.
- Probabilistic index models. Journal of Royal Statistical Society. Series B (Methodological), 74(4):623–671.
- Sparser ordinal regression models based on parametric and additive location-shift approaches. International Statistical Review, 90(2):306–327.
- Regularized regression for categorical data. Statistical Modelling, 16(3):161–200.
- Smoothing in ordinal regression: An application to sensory data. Stats, 4(3):616–633.
- Deep ordinal classification based on the proportional odds model. In From Bioinspired Systems and Biomedical Applications to Machine Learning, pages 441–451.
- Cumulative link models for deep ordinal classification. Neurocomputing, 401:48–58.
- Williams, R. (2006). Generalized ordered logit/partial proportional odds models for ordinal dependent variables. Stata Journal, 6(1):58–82.
- Williams, R. (2016). Understanding and interpreting generalized ordered logit models. The Journal of Mathematical Sociology, 40(1):7–20.
- Regularized ordinal regression and the ordinalNet R package. Journal of Statistical Software, 99(6):1–42.
- Deep lattice networks and partial monotonic functions. In Advances in Neural Information Processing Systems, volume 30.
- UCI Machine Learning Repository.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.