Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Prototype Rectification for Few-Shot Learning (1911.10713v4)

Published 25 Nov 2019 in cs.CV

Abstract: Few-shot learning requires to recognize novel classes with scarce labeled data. Prototypical network is useful in existing researches, however, training on narrow-size distribution of scarce data usually tends to get biased prototypes. In this paper, we figure out two key influencing factors of the process: the intra-class bias and the cross-class bias. We then propose a simple yet effective approach for prototype rectification in transductive setting. The approach utilizes label propagation to diminish the intra-class bias and feature shifting to diminish the cross-class bias. We also conduct theoretical analysis to derive its rationality as well as the lower bound of the performance. Effectiveness is shown on three few-shot benchmarks. Notably, our approach achieves state-of-the-art performance on both miniImageNet (70.31% on 1-shot and 81.89% on 5-shot) and tieredImageNet (78.74% on 1-shot and 86.92% on 5-shot).

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Jinlu Liu (17 papers)
  2. Liang Song (60 papers)
  3. Yongqiang Qin (6 papers)
Citations (229)

Summary

Prototype Rectification for Few-Shot Learning

In the paper "Prototype Rectification for Few-Shot Learning," the authors present a methodological advancement in few-shot learning by addressing the biases in prototype computation. Few-shot learning is known for dealing with novel classes using minimal labeled data. Traditional deep learning models tend to underperform in such scenarios due to a lack of sufficient data, leading researchers to explore novel approaches that can efficiently learn new concepts with limited examples.

Key Contributions

The research identifies two crucial biases in prototype-based few-shot learning methods: intra-class bias and cross-class bias. The intra-class bias arises due to inadequate data representing a single class, leading to a skewed prototype that does not accurately capture the class's central tendency. Cross-class bias occurs due to domain discrepancies between the training (support set) and the testing (query set) samples, which can mislead the classification of test instances.

The paper proposes a robust solution in the form of a bias diminishing module integrated within a cosine similarity-based prototypical network. This module operates in a transductive setting, enhancing the robustness and accuracy of prototype-based few-shot classification.

Methodological Innovation

The proposed approach involves two primary mechanisms:

  1. Intra-Class Bias Rectification: By utilizing a pseudo-labeling strategy, the method augments the support set with high-confidence predictions from the query set. This transductive approach amplifies the data available to form class prototypes, effectively reducing intra-class bias. Unlike simple averaging, the proposed method employs a weighted sum strategy for prototype computation to mitigate errors from misclassified samples.
  2. Cross-Class Bias Rectification: The method introduces a feature-shifting mechanism by incorporating a shifting term to query features, aligning them closer to the support set's distribution. This alignment reduces cross-class bias, ensuring that the distance between the domains (support and query sets) is minimized, which in turn improves classification accuracy.

Experimental Validation

The authors validate their approach on several benchmarks, including miniImageNet and tieredImageNet, achieving state-of-the-art performance. On miniImageNet, the proposed method achieved an accuracy of 70.31% for 1-shot and 81.89% for 5-shot learning tasks. Similar robust improvements were observed on tieredImageNet, with accuracies of 78.74% and 86.92% for 1-shot and 5-shot settings, respectively. These results highlight not only the effectiveness but also the practicality of the proposed approach.

Theoretical Insights

An in-depth theoretical analysis supports the empirical findings, demonstrating that the proposed methods effectively reduce the biases and improve performance lower bounds. The paper derives theoretical affirmations that link the rectified prototype computation mechanisms with improved accuracy, reinforcing its mathematical soundness.

Implications and Future Directions

The findings from this paper offer significant implications for the domain of few-shot learning:

  • Theoretical Implications: The clear articulation of bias factors in prototype-based learning enriches the understanding of model inefficiencies in low-data scenarios. Theoretically motivated solutions, such as those proposed, provide pathways to further explore the link between bias minimization and practical model improvements.
  • Practical Implications: The proposed solutions are not only mathematically grounded but also simple and easily integrable, making them appealing for real-world applications where data acquisition is costly and limited.

Looking forward, this research opens avenues for further innovation in few-shot learning. Future work may explore adaptive strategies for determining the number of pseudo-label samples per class or dynamically modifying the shifting term based on real-time feature distribution analysis. Additionally, extending this framework to other domains like semi-supervised learning or unsupervised domain adaptation could offer intriguing results.

This paper sets a solid foundation, motivating a deeper exploration of prototype rectification and bias minimization strategies to confront the challenges intrinsic to few-shot learning.