Scaling Up Semi-supervised Learning with Unconstrained Unlabelled Data (2306.01222v2)
Abstract: We propose UnMixMatch, a semi-supervised learning framework which can learn effective representations from unconstrained unlabelled data in order to scale up performance. Most existing semi-supervised methods rely on the assumption that labelled and unlabelled samples are drawn from the same distribution, which limits the potential for improvement through the use of free-living unlabeled data. Consequently, the generalizability and scalability of semi-supervised learning are often hindered by this assumption. Our method aims to overcome these constraints and effectively utilize unconstrained unlabelled data in semi-supervised learning. UnMixMatch consists of three main components: a supervised learner with hard augmentations that provides strong regularization, a contrastive consistency regularizer to learn underlying representations from the unlabelled data, and a self-supervised loss to enhance the representations that are learnt from the unlabelled data. We perform extensive experiments on 4 commonly used datasets and demonstrate superior performance over existing semi-supervised methods with a performance boost of 4.79%. Extensive ablation and sensitivity studies show the effectiveness and impact of each of the proposed components of our method.
- Vicreg: Variance-invariance-covariance regularization for self-supervised learning. In International Conference on Learning Representations.
- Remixmatch: Semi-Supervised learning with distribution matching and augmentation anchoring. In International Conference on Learning Representations.
- Mixmatch: A holistic approach to semi-supervised learning. In Advances in Neural Information Processing Systems.
- A simple framework for contrastive learning of visual representations. In International Conference on Machine Learning, 1597–1607.
- Exploring simple siamese representation learning. In Computer Vision and Pattern Recognition, 15750–15758.
- An analysis of single-layer networks in unsupervised feature learning. In International Conference on Artificial Intelligence and Statistics, 215–223.
- Randaugment: Practical automated data augmentation with a reduced search space. In Computer Vision and Pattern Recognition Workshops, 702–703.
- Imagenet: A large-scale hierarchical image database. In International Conference on Computer Vision, 248–255.
- Unsupervised representation learning by predicting image rotations. In International Conference on Learning Representations.
- Bootstrap your own latent-a new approach to self-supervised learning. In Advances in Neural Information Processing Systems, 21271–21284.
- Improving barely supervised learning by discriminating unlabeled samples with super-class. In Advances in Neural Information Processing Systems.
- Safe deep semi-supervised learning for unseen-class unlabeled data. In International Conference on Machine Learning, 3897–3906.
- Conmatch: Semi-supervised learning with confidence-guided consistency regularization. In European Conference on Computer Vision, 674–690.
- Learning multiple layers of features from tiny images. Technical report, University of Toronto.
- Contrastive regularization for semi-supervised learning. In Computer Vision and Pattern Recognition, 3911–3920.
- Lee, D.-H.; et al. 2013. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on Challenges in Representation Learning, ICML.
- Comatch: Semi-supervised learning with contrastive graph regularization. In International Conference on Computer Vision, 9475–9484.
- Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, (8): 1979–1993.
- Reading digits in natural images with unsupervised feature learning. In NeurIPS Workshop on Deep Learning and Unsupervised Feature Learning.
- Realistic evaluation of deep semi-supervised learning algorithms. In Advances in Neural Information Processing Systems.
- Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748.
- Impact of labelled set selection and supervision policies on semi-supervised learning. arXiv preprint arXiv:2211.14912.
- Openmatch: Open-set consistency regularization for semi-supervised learning with outliers. In Advances in Neural Information Processing Systems.
- Regularization with stochastic transformations and perturbations for deep semi-supervised learning. In Advances in Neural Information Processing Systems.
- Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Advances in Neural Information Processing Systems, 596–608.
- A realistic evaluation of semi-supervised learning for fine-grained classification. In Computer Vision and Pattern Recognition, 12966–12975.
- Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Advances in Neural Information Processing Systems.
- Unsupervised data augmentation for consistency training. In Advances in Neural Information Processing Systems, 6256–6268.
- Self-training with noisy student improves imagenet classification. In Computer Vision and Pattern Recognition, 10687–10698.
- Class-aware contrastive semi-supervised learning. In Computer Vision and Pattern Recognition, 14421–14430.
- Classification-reconstruction learning for open-set recognition. In Computer Vision and Pattern Recognition, 4016–4025.
- Multi-task curriculum framework for open-set semi-supervised learning. In European Conference on Computer Vision, 438–454.
- Wide residual networks. In British Machine Vision Conference.
- S4l: Self-supervised semi-supervised learning. In International Conference on Computer Vision, 1476–1485.
- Flexmatch: boosting semi-supervised learning with curriculum pseudo labeling. In Advances in Neural Information Processing Systems, 18408–18419.
- mixup: Beyond empirical risk minimization. In International Conference on Learning Representations.
- Colorful image colorization. In European Conference on Computer Vision, 649–666.
- SimMatch: Semi-supervised learning with similarity matching. In Computer Vision and Pattern Recognition, 14471–14481.
- Shuvendu Roy (18 papers)
- Ali Etemad (118 papers)