Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Scaling of Class-wise Training Losses for Post-hoc Calibration (2306.10989v1)

Published 19 Jun 2023 in cs.LG

Abstract: The class-wise training losses often diverge as a result of the various levels of intra-class and inter-class appearance variation, and we find that the diverging class-wise training losses cause the uncalibrated prediction with its reliability. To resolve the issue, we propose a new calibration method to synchronize the class-wise training losses. We design a new training loss to alleviate the variance of class-wise training losses by using multiple class-wise scaling factors. Since our framework can compensate the training losses of overfitted classes with those of under-fitted classes, the integrated training loss is preserved, preventing the performance drop even after the model calibration. Furthermore, our method can be easily employed in the post-hoc calibration methods, allowing us to use the pre-trained model as an initial model and reduce the additional computation for model calibration. We validate the proposed framework by employing it in the various post-hoc calibration methods, which generally improves calibration performance while preserving accuracy, and discover through the investigation that our approach performs well with unbalanced datasets and untuned hyperparameters.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. Bfgs optimization for faster and automated supervised learning. In International neural network conference, pp.  757–760. Springer, 1990.
  2. Panoptic-deeplab: A simple, strong, and fast baseline for bottom-up panoptic segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  12475–12485, 2020.
  3. Imagenet: A large-scale hierarchical image database. In IEEE conference on computer vision and pattern recognition, 2009.
  4. Foody, G. M. Impacts of sample design for validation data on the accuracy of feedforward neural network classification. Applied Sciences, 7(9):888, 2017.
  5. Network calibration by class-based temperature scaling. In 2021 29th European Signal Processing Conference (EUSIPCO). IEEE, 2021.
  6. Girshick, R. Fast r-cnn. In Proceedings of the IEEE international conference on computer vision, pp.  1440–1448, 2015.
  7. On calibration of modern neural networks. In International Conference on Machine Learning, 2017.
  8. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition, 2016.
  9. Deep networks with stochastic depth. In Leibe, B., Matas, J., Sebe, N., and Welling, M. (eds.), European Conference on Computer Vision, 2016.
  10. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
  11. Adam: A method for stochastic optimization. International Conference for Learning Representations, 2015.
  12. Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009.
  13. Imagenet classification with deep convolutional neural networks. Communications of the ACM, 2017.
  14. Beyond temperature scaling: Obtaining well-calibrated multiclass probabilities with dirichlet calibration. In Advances in Neural Information Processing Systems, 2019.
  15. Dynamic class queue for large scale face recognition in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  3763–3772, 2021.
  16. Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision, pp.  2980–2988, 2017.
  17. Meta-cal: Well-controlled post-hoc calibration by ranking. In International Conference on Machine Learning, 2021.
  18. Calibrating deep neural networks using focal loss. Advances in Neural Information Processing Systems, 33:15288–15299, 2020.
  19. When does label smoothing help? Advances in neural information processing systems, 32, 2019.
  20. Obtaining well calibrated probabilities using bayesian binning. In Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015.
  21. pyopt: a python-based object-oriented framework for nonlinear constrained optimization. Structural and Multidisciplinary Optimization, 45(1):101–118, 2012.
  22. Platt, J. et al. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in Large-Margin Classifiers, 1999.
  23. Intra order-preserving functions for calibration of multi-class neural networks. In Advances in Neural Information Processing Systems, 2020.
  24. Deep learning for self-driving cars: Chances and challenges. In Proceedings of the 1st International Workshop on Software Engineering for AI in Autonomous Systems, 2018.
  25. Very deep convolutional networks for large-scale image recognition. 2015.
  26. Long-tailed classification by keeping the good and removing the bad momentum causal effect. In NeurIPS, 2020.
  27. Parameterized temperature scaling for boosting the expressive power in post-hoc uncertainty calibration. In In European Conference on Computer Vision (ECCV), 2022.
  28. A comprehensive survey on convolutional neural network in medical image analysis. Multimedia Tools and Applications, 2020.
  29. Robustscanner: Dynamically enhancing positional clues for robust text recognition. In European Conference on Computer Vision.
  30. Wide residual networks. In Proceedings of the British Machine Vision Conference, 2016.
  31. Mix-n-match: Ensemble and compositional methods for uncertainty calibration in deep learning. In International Conference on Machine Learning, 2020.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Seungjin Jung (1 paper)
  2. Seungmo Seo (1 paper)
  3. Yonghyun Jeong (18 papers)
  4. Jongwon Choi (16 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.