Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improving Calibration for Long-Tailed Recognition (2104.00466v1)

Published 1 Apr 2021 in cs.CV

Abstract: Deep neural networks may perform poorly when training datasets are heavily class-imbalanced. Recently, two-stage methods decouple representation learning and classifier learning to improve performance. But there is still the vital issue of miscalibration. To address it, we design two methods to improve calibration and performance in such scenarios. Motivated by the fact that predicted probability distributions of classes are highly related to the numbers of class instances, we propose label-aware smoothing to deal with different degrees of over-confidence for classes and improve classifier learning. For dataset bias between these two stages due to different samplers, we further propose shifted batch normalization in the decoupling framework. Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets, including CIFAR-10-LT, CIFAR-100-LT, ImageNet-LT, Places-LT, and iNaturalist 2018. Code will be available at https://github.com/Jia-Research-Lab/MiSLAS.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Zhisheng Zhong (20 papers)
  2. Jiequan Cui (22 papers)
  3. Shu Liu (146 papers)
  4. Jiaya Jia (162 papers)
Citations (267)

Summary

Improving Calibration for Long-Tailed Recognition

The paper, "Improving Calibration for Long-Tailed Recognition," by Zhisheng Zhong et al., investigates the challenges deep neural networks face when dealing with long-tailed datasets, which are characterized by significant class imbalances. The authors propose novel methods to enhance calibration and recognition performance in such complex scenarios.

A core problem addressed by the paper is the miscalibration of neural networks trained on long-tailed distributions. Calibration, in this context, refers to the alignment between the predicted probabilities and the actual likelihood of those predictions being correct. Previous two-stage approaches to long-tailed recognition have decoupled representation and classifier learning to mitigate performance issues but continue to suffer from miscalibration and over-confidence.

To tackle these challenges, the authors propose two key enhancements: label-aware smoothing and shifted batch normalization within a decoupled learning framework. Label-aware smoothing adjusts the predicted class probabilities based on the number of instances per class, effectively reducing over-confidence in head classes with abundant data. This is achieved by dynamically tuning label smoothing parameters according to class frequencies. Shifted batch normalization addresses dataset biases between different sampling strategies employed in the two-stage framework, by updating running statistics while keeping linear transformation parameters fixed during the classifier retraining phase.

The paper presents compelling empirical results, establishing new performance benchmarks across several long-tailed datasets, including CIFAR-10-LT, CIFAR-100-LT, ImageNet-LT, Places-LT, and iNaturalist 2018. These experiments showcase the proposed methods' ability to achieve better accuracy and calibration compared to existing approaches, such as deferred re-weighting (DRW) and bilateral branch networks (BBN).

Furthermore, the paper's implications extend beyond improved recognition performance. By mitigating over-confidence and enhancing calibration, the methods contribute to the development of more reliable machine learning models that can make well-calibrated predictions about less-represented classes. This advancement is particularly relevant in applications requiring high confidence in model decisions across diverse and imbalanced real-world scenarios.

Future work could explore integrating these methods with other calibration-enhancing techniques or applying them to different domains. Additionally, further investigation into adaptive forms of label-aware smoothing and the quantification of dataset bias in more complex settings could yield valuable insights.

In summary, this paper provides a robust approach to addressing the twin challenges of miscalibration and dataset bias in long-tailed recognition, offering a significant step forward in developing reliable and accurate neural networks for real-world applications.

Github Logo Streamline Icon: https://streamlinehq.com