Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

AggMatch: Aggregating Pseudo Labels for Semi-Supervised Learning (2201.10444v1)

Published 25 Jan 2022 in cs.LG and cs.CV

Abstract: Semi-supervised learning (SSL) has recently proven to be an effective paradigm for leveraging a huge amount of unlabeled data while mitigating the reliance on large labeled data. Conventional methods focused on extracting a pseudo label from individual unlabeled data sample and thus they mostly struggled to handle inaccurate or noisy pseudo labels, which degenerate performance. In this paper, we address this limitation with a novel SSL framework for aggregating pseudo labels, called AggMatch, which refines initial pseudo labels by using different confident instances. Specifically, we introduce an aggregation module for consistency regularization framework that aggregates the initial pseudo labels based on the similarity between the instances. To enlarge the aggregation candidates beyond the mini-batch, we present a class-balanced confidence-aware queue built with the momentum model, encouraging to provide more stable and consistent aggregation. We also propose a novel uncertainty-based confidence measure for the pseudo label by considering the consensus among multiple hypotheses with different subsets of the queue. We conduct experiments to demonstrate the effectiveness of AggMatch over the latest methods on standard benchmarks and provide extensive analyses.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Jiwon Kim (50 papers)
  2. Kwangrok Ryoo (8 papers)
  3. Gyuseong Lee (11 papers)
  4. Seokju Cho (19 papers)
  5. Junyoung Seo (14 papers)
  6. Daehwan Kim (9 papers)
  7. Hansang Cho (8 papers)
  8. Seungryong Kim (103 papers)
Citations (1)