Scaling of Class-wise Training Losses for Post-hoc Calibration (2306.10989v1)
Abstract: The class-wise training losses often diverge as a result of the various levels of intra-class and inter-class appearance variation, and we find that the diverging class-wise training losses cause the uncalibrated prediction with its reliability. To resolve the issue, we propose a new calibration method to synchronize the class-wise training losses. We design a new training loss to alleviate the variance of class-wise training losses by using multiple class-wise scaling factors. Since our framework can compensate the training losses of overfitted classes with those of under-fitted classes, the integrated training loss is preserved, preventing the performance drop even after the model calibration. Furthermore, our method can be easily employed in the post-hoc calibration methods, allowing us to use the pre-trained model as an initial model and reduce the additional computation for model calibration. We validate the proposed framework by employing it in the various post-hoc calibration methods, which generally improves calibration performance while preserving accuracy, and discover through the investigation that our approach performs well with unbalanced datasets and untuned hyperparameters.
- Bfgs optimization for faster and automated supervised learning. In International neural network conference, pp. 757–760. Springer, 1990.
- Panoptic-deeplab: A simple, strong, and fast baseline for bottom-up panoptic segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 12475–12485, 2020.
- Imagenet: A large-scale hierarchical image database. In IEEE conference on computer vision and pattern recognition, 2009.
- Foody, G. M. Impacts of sample design for validation data on the accuracy of feedforward neural network classification. Applied Sciences, 7(9):888, 2017.
- Network calibration by class-based temperature scaling. In 2021 29th European Signal Processing Conference (EUSIPCO). IEEE, 2021.
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE international conference on computer vision, pp. 1440–1448, 2015.
- On calibration of modern neural networks. In International Conference on Machine Learning, 2017.
- Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition, 2016.
- Deep networks with stochastic depth. In Leibe, B., Matas, J., Sebe, N., and Welling, M. (eds.), European Conference on Computer Vision, 2016.
- Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
- Adam: A method for stochastic optimization. International Conference for Learning Representations, 2015.
- Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009.
- Imagenet classification with deep convolutional neural networks. Communications of the ACM, 2017.
- Beyond temperature scaling: Obtaining well-calibrated multiclass probabilities with dirichlet calibration. In Advances in Neural Information Processing Systems, 2019.
- Dynamic class queue for large scale face recognition in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3763–3772, 2021.
- Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision, pp. 2980–2988, 2017.
- Meta-cal: Well-controlled post-hoc calibration by ranking. In International Conference on Machine Learning, 2021.
- Calibrating deep neural networks using focal loss. Advances in Neural Information Processing Systems, 33:15288–15299, 2020.
- When does label smoothing help? Advances in neural information processing systems, 32, 2019.
- Obtaining well calibrated probabilities using bayesian binning. In Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015.
- pyopt: a python-based object-oriented framework for nonlinear constrained optimization. Structural and Multidisciplinary Optimization, 45(1):101–118, 2012.
- Platt, J. et al. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in Large-Margin Classifiers, 1999.
- Intra order-preserving functions for calibration of multi-class neural networks. In Advances in Neural Information Processing Systems, 2020.
- Deep learning for self-driving cars: Chances and challenges. In Proceedings of the 1st International Workshop on Software Engineering for AI in Autonomous Systems, 2018.
- Very deep convolutional networks for large-scale image recognition. 2015.
- Long-tailed classification by keeping the good and removing the bad momentum causal effect. In NeurIPS, 2020.
- Parameterized temperature scaling for boosting the expressive power in post-hoc uncertainty calibration. In In European Conference on Computer Vision (ECCV), 2022.
- A comprehensive survey on convolutional neural network in medical image analysis. Multimedia Tools and Applications, 2020.
- Robustscanner: Dynamically enhancing positional clues for robust text recognition. In European Conference on Computer Vision.
- Wide residual networks. In Proceedings of the British Machine Vision Conference, 2016.
- Mix-n-match: Ensemble and compositional methods for uncertainty calibration in deep learning. In International Conference on Machine Learning, 2020.
- Seungjin Jung (1 paper)
- Seungmo Seo (1 paper)
- Yonghyun Jeong (18 papers)
- Jongwon Choi (16 papers)