Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Generative Semi-supervised Learning with Meta-Optimized Synthetic Samples (2309.16143v1)

Published 28 Sep 2023 in cs.LG, cs.AI, cs.CV, and stat.ML

Abstract: Semi-supervised learning (SSL) is a promising approach for training deep classification models using labeled and unlabeled datasets. However, existing SSL methods rely on a large unlabeled dataset, which may not always be available in many real-world applications due to legal constraints (e.g., GDPR). In this paper, we investigate the research question: Can we train SSL models without real unlabeled datasets? Instead of using real unlabeled datasets, we propose an SSL method using synthetic datasets generated from generative foundation models trained on datasets containing millions of samples in diverse domains (e.g., ImageNet). Our main concepts are identifying synthetic samples that emulate unlabeled samples from generative foundation models and training classifiers using these synthetic samples. To achieve this, our method is formulated as an alternating optimization problem: (i) meta-learning of generative foundation models and (ii) SSL of classifiers using real labeled and synthetic unlabeled samples. For (i), we propose a meta-learning objective that optimizes latent variables to generate samples that resemble real labeled samples and minimize the validation loss. For (ii), we propose a simple unsupervised loss function that regularizes the feature extractors of classifiers to maximize the performance improvement obtained from synthetic samples. We confirm that our method outperforms baselines using generative foundation models on SSL. We also demonstrate that our methods outperform SSL using real unlabeled datasets in scenarios with extremely small amounts of labeled datasets. This suggests that synthetic samples have the potential to provide improvement gains more efficiently than real unlabeled data.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Learning with pseudo-ensembles. In Advances in neural information processing systems, 2014.
  2. Large scale GAN training for high fidelity natural image synthesis. In International Conference on Learning Representations, 2019.
  3. Semi-supervised Learning. MIT Press, 2006.
  4. Describing textures in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2014.
  5. Randaugment: Practical automated data augmentation with a reduced search space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pages 702–703, 2020.
  6. Good semi-supervised learning that requires a bad gan. Advances in neural information processing systems, 30, 2017.
  7. Universum gans: Improving gans through contradictions. arXiv preprint arXiv:2106.09946, 2021.
  8. Model-agnostic meta-learning for fast adaptation of deep networks. In International conference on machine learning, 2017.
  9. Ross Girshick. Fast r-cnn. In International Conference on Computer Vision, 2015.
  10. Generative adversarial nets. In Advances in Neural Information Processing Systems 27, 2014.
  11. Semi-supervised learning by entropy minimization. In Advances in Neural Information Processing Systems, 2005.
  12. Low-shot visual recognition by shrinking and hallucinating features. In Proceedings of the IEEE international conference on computer vision, pages 3018–3027, 2017.
  13. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016.
  14. Is synthetic data from generative models ready for image recognition? 2023.
  15. Categorical reparameterization with gumbel-softmax. In International Conference on Learning Representation, 2017.
  16. Training generative adversarial networks with limited data. In Advances in Neural Information Processing Systems, 2020.
  17. 3d object representations for fine-grained categorization. In 4th International IEEE Workshop on 3D Representation and Recognition, Sydney, Australia, 2013.
  18. Temporal ensembling for semi-supervised learning. In International Conference on Learning Representations, 2016.
  19. Dong-Hyun Lee et al. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, ICML, 2013.
  20. Model adaptation: Unsupervised domain adaptation without source data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
  21. On the privacy properties of gan-generated samples. In International Conference on Artificial Intelligence and Statistics, pages 1522–1530. PMLR, 2021.
  22. Fine-grained visual classification of aircraft. arXiv, 2013.
  23. M-E. Nilsback and A. Zisserman. Automated flower classification over a large number of classes. In Proceedings of the Indian Conference on Computer Vision, Graphics and Image Processing, 2008.
  24. Conditional image synthesis with auxiliary classifier gans. In International Conference on Machine Learning, 2017.
  25. Cats and dogs. In IEEE Conference on Computer Vision and Pattern Recognition, 2012.
  26. Kornia: an open source differentiable computer vision library for pytorch. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2020.
  27. Regularization with stochastic transformations and perturbations for deep semi-supervised learning. In Advances in neural information processing systems, 2016.
  28. Generate to adapt: Aligning domains using generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018.
  29. Stylegan-xl: Scaling stylegan to large diverse datasets. In ACM SIGGRAPH 2022 conference proceedings, pages 1–10, 2022.
  30. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Advances in Neural Information Processing Systems, 2020.
  31. Synthetic data, real errors: how (not) to publish and use synthetic data. In International Conference on Machine Learning, 2023.
  32. Freematch: Self-adaptive thresholding for semi-supervised learning. In International Conference on Learning Representations, 2023.
  33. Caltech-UCSD Birds 200. Technical report, California Institute of Technology, 2010.
  34. Unsupervised data augmentation for consistency training. In Advances in Neural Information Processing Systems, 2020.
  35. Effective data augmentation with multi-domain learning gans. In Proceedings of the AAAI Conference on Artificial Intelligence, 2020.
  36. Transfer learning with pre-trained conditional generative models. arXiv preprint arXiv:2204.12833, 2022.
  37. Regularizing neural networks with meta-learning generative models. In Advances in Neural Information Processing Systems, 2023.
  38. Data augmentation using conditional generative adversarial networks for leaf counting in arabidopsis plants. In British Machine Vision Conference, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Shin'ya Yamaguchi (24 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.