Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Mining Latent Classes for Few-shot Segmentation (2103.15402v3)

Published 29 Mar 2021 in cs.CV

Abstract: Few-shot segmentation (FSS) aims to segment unseen classes given only a few annotated samples. Existing methods suffer the problem of feature undermining, i.e. potential novel classes are treated as background during training phase. Our method aims to alleviate this problem and enhance the feature embedding on latent novel classes. In our work, we propose a novel joint-training framework. Based on conventional episodic training on support-query pairs, we add an additional mining branch that exploits latent novel classes via transferable sub-clusters, and a new rectification technique on both background and foreground categories to enforce more stable prototypes. Over and above that, our transferable sub-cluster has the ability to leverage extra unlabeled data for further feature enhancement. Extensive experiments on two FSS benchmarks demonstrate that our method outperforms previous state-of-the-art by a large margin of 3.7% mIOU on PASCAL-5i and 7.0% mIOU on COCO-20i at the cost of 74% fewer parameters and 2.5x faster inference speed. The source code is available at https://github.com/LiheYoung/MiningFSS.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Lihe Yang (12 papers)
  2. Wei Zhuo (24 papers)
  3. Lei Qi (84 papers)
  4. Yinghuan Shi (79 papers)
  5. Yang Gao (761 papers)
Citations (117)

Summary

Mining Latent Classes for Few-shot Segmentation

The paper "Mining Latent Classes for Few-shot Segmentation" introduces an innovative framework addressing the problem of few-shot segmentation (FSS), which involves segmenting unseen classes with limited annotated data. Current methods in FSS often misclassify potential novel classes as background categories during training, leading to feature undermining. This research proposes a joint-training framework incorporating a novel mining branch aimed at identifying latent novel classes via transferable sub-clusters, alongside a rectification technique to stabilize class prototypes. The efficacy of this approach is demonstrated through substantial improvements in performance benchmarks, as well as reductions in model size and inference time.

Key Contributions

  1. Joint-training Framework: This method incorporates episodic training on support-query pairs with an auxiliary mining branch. The mining branch enhances the feature embeddings by utilizing latent novel classes identified through transferable sub-clusters, which are derived from the base classes. This dual strategy fosters the ability to generalize effectively to unseen classes without further training or fine-tuning, addressing the critical issue of feature undermining.
  2. Rectification of Prototypes: The paper introduces a prototype rectification technique to mitigate prototype bias, which is a common issue in FSS due to limited support samples. This is achieved by refining both foreground and background prototypes. A global background prototype is maintained and updated as a moving average of all training set backgrounds, broadening the context. For foreground classes, region-level prototypes from additional samples are integrated.
  3. Empirical Validation: The framework was empirically validated with experiments on PASCAL-5i^i and COCO-20i^i datasets, where it outperformed existing state-of-the-art models, marking improvements of 3.7% mIOU on PASCAL-5i^i and 7.0% mIOU on COCO-20i^i. Remarkably, these enhancements were achieved with 74% fewer parameters and increased inference speed by 2.5x.
  4. Exploitation of Unlabeled Data: Beyond the improvements derived from the labeled data, the proposed methodology can integrate additional unlabeled data, further enhancing performance. This is a significant step towards more realistic few-shot learning settings where class labels are frequently sparse or entirely absent.

Implications

The presented paper provides both theoretical advancements and practical improvements within the field of few-shot segmentation. The concept of mining latent novel classes through pseudo labeling introduces an innovative approach to leverage unannotated or partially annotated data, aligning well with the constant influx of unlabeled data in real-world applications. Practically, the reduction in model complexity and enhanced inference speed can facilitate the deployment of segmentation models in resource-constrained environments, such as mobile devices and edge computing.

Future Directions

Future work could explore further automation in the selection of sub-clusters, potentially utilizing more sophisticated clustering algorithms or data-driven techniques to adaptively determine the number and characteristics of sub-clusters. Additionally, with advancements in other areas of meta-learning and semi-supervised learning, the potential integration of cross-domain knowledge could further extend the applicability of FSS methods.

In conclusion, this research significantly contributes to the domain of few-shot learning by introducing a method that not only enhances performance metrics but is designed with efficiency and scalability in mind, making it well-suited for the increasingly demanding applications of computer vision technology.

Github Logo Streamline Icon: https://streamlinehq.com