Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
175 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Uncertainty Voting Ensemble for Imbalanced Deep Regression (2305.15178v4)

Published 24 May 2023 in cs.LG

Abstract: Data imbalance is ubiquitous when applying machine learning to real-world problems, particularly regression problems. If training data are imbalanced, the learning is dominated by the densely covered regions of the target distribution and the learned regressor tends to exhibit poor performance in sparsely covered regions. Beyond standard measures like oversampling or reweighting, there are two main approaches to handling learning from imbalanced data. For regression, recent work leverages the continuity of the distribution, while for classification, the trend has been to use ensemble methods, allowing some members to specialize in predictions for sparser regions. In our method, named UVOTE, we integrate recent advances in probabilistic deep learning with an ensemble approach for imbalanced regression. We replace traditional regression losses with negative log-likelihood, which also predicts sample-wise aleatoric uncertainty. Our experiments show that this loss function handles imbalance better. Additionally, we use the predicted aleatoric uncertainty values to fuse the predictions of different expert models in the ensemble, eliminating the need for a separate aggregation module. We compare our method with existing alternatives on multiple public benchmarks and show that UVOTE consistently outperforms the prior art, while at the same time producing better-calibrated uncertainty estimates. Our code is available at https://github.com/SherryJYC/UVOTE.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Country-wide retrieval of forest structure from optical and SAR satellite imagery with deep ensembles. ISPRS Journal of Photogrammetry and Remote Sensing, 195:269–286, 2023.
  2. SMOGN: a pre-processing approach for imbalanced regression. In International Workshop on Learning with Imbalanced Domains: Theory and Applications, pages 36–50, 2017.
  3. Semantic textual similarity-multilingual and cross-lingual focused evaluation. In International Workshop on Semantic Evaluation (SemEval), 2017.
  4. SMOTE: synthetic minority over-sampling technique. Journal of Artificial Intelligence Research, 16:321–357, 2002.
  5. Class-balanced loss based on effective number of samples. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9268–9277, 2019.
  6. RankSim: Ranking similarity regularization for deep imbalanced regression. arXiv preprint arXiv:2205.15236, 2022.
  7. Learning from imbalanced data. IEEE Transactions on Knowledge and Data Engineering, 2009.
  8. Revisiting single image depth estimation: Toward higher resolution maps with accurate object boundaries. In IEEE Winter Conference on Applications of Computer Vision (WACV), pages 1043–1051. IEEE, 2019.
  9. What uncertainties do we need in bayesian deep learning for computer vision? Advances in Neural Information Processing Systems (NeurIPS), 30, 2017.
  10. A high-resolution canopy height model of the earth. Nature Ecology & Evolution, pages 1–12, 2023.
  11. Deepti: Deep-learning-based tropical cyclone intensity estimation system. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 13:4271–4281, 2020.
  12. AgeDB: the first manually collected, in-the-wild age database. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 51–59, 2017.
  13. Pytorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems (NeurIPS), 32, 2019.
  14. Balanced mse for imbalanced visual regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7926–7935, 2022.
  15. Deep expectation of real and apparent age from a single image without facial landmarks. International Journal of Computer Vision, 126(2-4):144–157, 2018.
  16. Indoor segmentation and support inference from rgbd images. In ECCV, 2012.
  17. Density-based weighting for imbalanced regression. Machine Learning, 110:2187–2211, 2021.
  18. GLUE: A multi-task benchmark and analysis platform for natural language understanding. arXiv preprint arXiv:1804.07461, 2018.
  19. Long-tailed recognition by routing diverse distribution-aware experts. arXiv preprint arXiv:2010.01809, 2020.
  20. Delving into deep imbalanced regression. In International Conference on Machine Learning (ICML), pages 11842–11851, 2021.
  21. Robustness via cross-domain ensembles. In IEEE/CVF International Conference on Computer Vision (ICCV), pages 12189–12199, 2021.
  22. mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412, 2017.
  23. Self-supervised aggregation of diverse experts for test-agnostic long-tailed recognition. arXiv preprint arXiv:2107.09249, 2021.
  24. BBN: Bilateral-branch network with cumulative learning for long-tailed visual recognition. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9719–9728, 2020.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets