Imbalanced Deep Learning by Minority Class Incremental Rectification
Imbalanced data distribution poses a significant challenge in deep learning, particularly when the dataset includes a disproportionally large number of samples for certain classes compared to others. This paper by Qi Dong, Shaogang Gong, and Xiatian Zhu presents a novel approach to tackling this problem through a method termed as "Class Rectification Loss" (CRL), aimed at improving the recognition of minority classes within imbalanced datasets.
The primary focus of the paper is on enhancing deep learning models to perform better on severely imbalanced data, which is common in many real-world applications. Existing approaches commonly rely on pre-processing techniques like data re-sampling or algorithmic modifications like cost-sensitive learning. However, these methods have limitations, such as potential overfitting with over-sampling or loss of valuable information with down-sampling.
The authors propose a batch-wise incremental approach where minority class rectification occurs through hard sample mining. This process involves selecting the hardest samples from the minority classes in each mini-batch during training. The effectiveness of this method rests on a technique that incrementally weeds out model learning bias towards majority classes by focusing computational resources on the harder aspects of minority class boundary discovery and delineation.
The authors introduce a Class Rectification Loss (CRL) function that can be integrated into existing deep network architectures. CRL works in tandem with the standard cross-entropy loss, interrogating models into learning more informative representations of minority class features. Crucially, it integrates into the batch-wise processing paradigm of modern deep networks, enabling scalability to large datasets without global data pre-processing burdens or assumptions.
Three core contributions of the paper stand out:
- It presents a scalable solution for class imbalanced deep learning by leveraging batch-wise optimisation.
- It innovatively includes hard sample mining within mini-batch training to rectify model bias, making significant headway over existing sampling or cost-sensitive methods.
- It demonstrates the CRL's applicability and improvements across datasets with different nature and degree of imbalance, showing significant advancements in facial and clothing attribute recognition benchmarks.
The empirical evaluation spans multiple datasets, including significantly imbalanced person attribute datasets like CelebA and X-Domain, as well as a balanced dataset, CIFAR-100, to underline the versatility and effectiveness of their approach. The experiments show notable improvements in class-balanced accuracy, demonstrating the higher sensitivity and precision of CRL-equipped models for minority classes compared to traditional methods and state-of-the-art imbalanced learning techniques.
The implications of these findings are both practical and theoretical. Practically, the proposed method can enhance the deployment of deep learning models in domains laden with naturally imbalanced data distributions, such as healthcare diagnostics or rare event detection. Theoretically, these results underscore the potential for better embedding learning that remains adaptive even under vastly skewed data conditions. CRL's design leveraging hard sample mining also presents an intriguing pathway for future research, particularly in the realms of online learning and active learning where data imbalance is inherently severe.
This paper provides foundational insights into the mechanics and potentials of addressing class imbalance through deep learning, emphasizing the importance of rectifying sample representation progressively and locally in model training. Future research directions illuminated by this work include extending CRL frameworks to integrate directly with online deep learning systems, enabling real-time adaptive learning with imbalanced streams of data.