Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
140 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Parametric Classification for Generalized Category Discovery: A Baseline Study (2211.11727v4)

Published 21 Nov 2022 in cs.CV and cs.LG

Abstract: Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples. Previous studies argued that parametric classifiers are prone to overfitting to seen categories, and endorsed using a non-parametric classifier formed with semi-supervised k-means. However, in this study, we investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem. We demonstrate that two prediction biases exist: the classifier tends to predict seen classes more often, and produces an imbalanced distribution across seen and novel categories. Based on these findings, we propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers. We hope the investigation and proposed simple framework can serve as a strong baseline to facilitate future studies in this field. Our code is available at: https://github.com/CVMI-Lab/SimGCD.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (57)
  1. k-means++: the advantages of careful seeding. In ACM-SIAM Symposium on Discrete Algorithms, 2007.
  2. Self-labelling via simultaneous clustering and representation learning. In ICLR, 2020.
  3. Masked siamese networks for label-efficient learning. In ECCV, 2022.
  4. Semi-supervised learning of visual features by non-parametrically predicting view assignments with support samples. In ICCV, 2021.
  5. Mixmatch: A holistic approach to semi-supervised learning. In NeurIPS, 2019.
  6. Open-world semi-supervised learning. In ICLR, 2022.
  7. Deep Clustering for Unsupervised Learning of Visual Features. In ECCV, 2018.
  8. Unsupervised learning of visual features by contrasting cluster assignments. In nips, 2020.
  9. Emerging properties in self-supervised vision transformers. In ICCV, 2021.
  10. A simple framework for contrastive learning of visual representations. In ICML, 2020.
  11. Semi-supervised learning under class distribution mismatch. In AAAI, 2020.
  12. Picie: Unsupervised semantic segmentation using invariance and equivariance in clustering. In CVPR, 2021.
  13. Discriminability-transferability trade-off: An information-theoretic perspective. In ECCV, 2022.
  14. Imagenet: A large-scale hierarchical image database. In CVPR, 2009.
  15. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021.
  16. Xcon: Learning with experts for fine-grained category discovery. In BMVC, 2022.
  17. A unified objective for novel class discovery. In ICCV, 2021.
  18. Unsupervised representation learning by predicting image rotations. In ICLR, 2018.
  19. Bootstrap your own latent-a new approach to self-supervised learning. In NeurIPS, 2020.
  20. Safe deep semi-supervised learning for unseen-class unlabeled data. In ICML, 2020.
  21. Automatically discovering and learning new visual categories with ranking statistics. In ICLR, 2020.
  22. Autonovel: Automatically discovering and learning novel visual categories. IEEE TPAMI, 2021.
  23. Learning to discover novel visual categories via deep transfer clustering. In ICCV, 2019.
  24. Momentum contrast for unsupervised visual representation learning. In CVPR, 2020.
  25. Deep residual learning for image recognition. In CVPR, 2016.
  26. Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 2019.
  27. Supervised contrastive learning. In NeurIPS, 2020.
  28. 3d object representations for fine-grained categorization. In ICCV Workshops, 2013.
  29. Learning multiple layers of features from tiny images. Technical Report, 2009.
  30. James MacQueen. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, 1967.
  31. Fine-grained visual classification of aircraft. arXiv preprint arXiv:1306.5151, 2013.
  32. Long-tail learning via logit adjustment. In ICLR, 2021.
  33. Realistic evaluation of deep semi-supervised learning algorithms. In NeurIPS, 2018.
  34. Semi-supervised learning with scarce annotations. In CVPR Workshops, 2020.
  35. Balanced meta-softmax for long-tailed visual recognition. In NeurIPS, 2020.
  36. Class-incremental novel class discovery. In ECCV, 2022.
  37. Openmatch: Open-set semi-supervised learning with open-set consistency regularization. In NeurIPS, 2021.
  38. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In NeurIPS, 2020.
  39. Opencon: Open-world contrastive learning. TMLR, 2023.
  40. The herbarium challenge 2019 dataset. arXiv preprint arXiv:1906.05372, 2019.
  41. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In NeurIPS, 2017.
  42. Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. JMLR, 2008.
  43. Generalized category discovery. In CVPR, 2022.
  44. Open-set recognition: A good closed-set classifier is all you need? In ICLR, 2022.
  45. Normface: L2 hypersphere embedding for face verification. In ACM MM, 2017.
  46. Long-tailed recognition by routing diverse distribution-aware experts. In ICLR, 2021.
  47. Debiased learning from naturally imbalanced pseudo-labels. In CVPR, 2022.
  48. Caltech-ucsd birds 200. Technical Report CNS-TR-201, Caltech, 2010.
  49. Self-supervised visual representation learning with semantic grouping. In NeurIPS, 2022.
  50. Multi-task curriculum framework for open-set semi-supervised learning. In ECCV, 2020.
  51. S4l: Self-supervised semi-supervised learning. In ICCV, 2019.
  52. Novel visual category discovery with dual ranking statistics and mutual knowledge distillation. In NeurIPS, 2021.
  53. Incremental generalized category discovery. In ICCV, 2023.
  54. Distilling visual priors from self-supervised learning. In ECCV Workshops, 2020.
  55. Learning semi-supervised gaussian mixture models for generalized category discovery. In ICCV, 2023.
  56. Neighborhood contrastive learning for novel class discovery. In CVPR, 2021.
  57. Openmix: Reviving known knowledge for discovering novel visual categories in an open world. In CVPR, 2021.
Citations (47)

Summary

  • The paper demonstrates that reliable pseudo labeling and entropy regularization can overcome overfitting, making parametric classifiers effective for GCD.
  • It introduces a self-distillation framework that refines predictions from augmented data to mitigate bias toward seen classes.
  • The study establishes a robust baseline by achieving state-of-the-art performance on benchmarks, challenging traditional non-parametric methods.

Overview of the Paper: Parametric Classification for Generalized Category Discovery

The paper under discussion, "Parametric Classification for Generalized Category Discovery: A Baseline Study," reinvestigates the efficacy of parametric classifiers for the generalized category discovery (GCD) task. This research landscapes an important point—contrary to previous assertions that parametric classifiers overfit seen categories, making them inadequate for novel category recognition, this paper illustrates that reliable pseudo labeling coupled with appropriate regularization can notably improve these classifiers' performance.

GCD is a variant within the supervised learning paradigm where the objective is to identify novel categories in unlabeled datasets, leveraging knowledge from labeled samples. Traditional methodologies advocated non-parametric classifiers based on semi-supervised kk-means due to their perceived robustness. However, this paper re-evaluates parametric methods’ potential, offering a fresh baseline solution bolstered by entropy regularization and an innovative self-distillation framework, attributing former failures to unreliable pseudo-label generation.

Key Contributions

  1. Revisiting Parametric Classifiers: The authors delve into parametric classifiers' perceived inadequacies, highlighting that their ineffectiveness is primarily due to unreliable pseudo labels rather than the classifier design itself. By addressing the biases in predictions—such as the tendency to overpredict seen classes or generate imbalanced distributions—through entropy regularization, they effectively mitigate previous overfitting concerns.
  2. Implementation of Self-distillation Framework: They propose a self-distillation technique for pseudo labeling, which significantly enhances classifier robustness against category biases. The classifier, during training, utilizes predictions from alternative data augmentations, refined through entropy maximization, to produce pseudo labels that guide parameter learning and help in better class separability.
  3. Advanced Baseline Establishment: By integrating entropy regularization and self-distillation into the training paradigm, a robust baseline for GCD is developed. The paper demonstrates how their parametric framework achieves state-of-the-art results on several benchmarks, showcasing resilience to the number of unknown class categories, a scenario often encountered in real-world deployments.

Implications

The implications of this research are profound, reshaping the belief about parametric classifiers' role in GCD tasks. By addressing prediction biases and optimizing supervision through pseudo labels, this paper suggests that parametric methods can achieve parity, if not superiority, with non-parametric techniques like kk-means. This examination highlights the potential for these classifiers to reduce computational overhead while maintaining precision in large-scale operations.

Speculations on Future Developments

The research invites several avenues for future exploration, such as enhancing entropy regularization to adapt dynamically to data variations and further refinement of self-distillation processes to make them even more reflective of true class distributions. Moreover, potential integration with unsupervised representation learning strategies could yield even more versatile, scalable solutions for GCD. Long-term, such advancements could streamline frameworks tackling open-set recognition problems across domains beyond traditional image datasets, heralding versatile parametric classifiers as a critical component in the evolving AI landscape.

In conclusion, this paper effectively challenges and redefines previous suppositions about parametric classifiers for GCD, offering a robust baseline poised to influence future research trajectories significantly. The careful methodological enhancements, combined with rigorous quantitative demonstrations, underscore its contribution to advancing the state-of-the-art in generalized category discovery.