Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-Class Classification from Noisy-Similarity-Labeled Data (2002.06508v1)

Published 16 Feb 2020 in cs.LG and stat.ML

Abstract: A similarity label indicates whether two instances belong to the same class while a class label shows the class of the instance. Without class labels, a multi-class classifier could be learned from similarity-labeled pairwise data by meta classification learning. However, since the similarity label is less informative than the class label, it is more likely to be noisy. Deep neural networks can easily remember noisy data, leading to overfitting in classification. In this paper, we propose a method for learning from only noisy-similarity-labeled data. Specifically, to model the noise, we employ a noise transition matrix to bridge the class-posterior probability between clean and noisy data. We further estimate the transition matrix from only noisy data and build a novel learning system to learn a classifier which can assign noise-free class labels for instances. Moreover, we theoretically justify how our proposed method generalizes for learning classifiers. Experimental results demonstrate the superiority of the proposed method over the state-of-the-art method on benchmark-simulated and real-world noisy-label datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Songhua Wu (2 papers)
  2. Xiaobo Xia (44 papers)
  3. Tongliang Liu (251 papers)
  4. Bo Han (283 papers)
  5. Mingming Gong (135 papers)
  6. Nannan Wang (106 papers)
  7. Haifeng Liu (56 papers)
  8. Gang Niu (125 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.