Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Phased Data Augmentation for Training a Likelihood-Based Generative Model with Limited Data (2305.12681v2)

Published 22 May 2023 in cs.CV, cs.LG, and eess.IV

Abstract: Generative models excel in creating realistic images, yet their dependency on extensive datasets for training presents significant challenges, especially in domains where data collection is costly or challenging. Current data-efficient methods largely focus on GAN architectures, leaving a gap in training other types of generative models. Our study introduces "phased data augmentation" as a novel technique that addresses this gap by optimizing training in limited data scenarios without altering the inherent data distribution. By limiting the augmentation intensity throughout the learning phases, our method enhances the model's ability to learn from limited data, thus maintaining fidelity. Applied to a model integrating PixelCNNs with VQ-VAE-2, our approach demonstrates superior performance in both quantitative and qualitative evaluations across diverse datasets. This represents an important step forward in the efficient training of likelihood-based models, extending the usefulness of data augmentation techniques beyond just GANs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. N.T. Tran, V.H. Tran, N.B. Nguyen, T.K. Nguyen, and N.M. Cheung, “On data augmentation for gan training,” IEEE Transactions on Image Processing, vol.30, pp.1882–1897, 2021.
  2. Z. Wang, Z. Dai, B. Póczos, and J. Carbonell, “Characterizing and avoiding negative transfer,” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.11293–11302, 2019.
  3. S. Zhao, Z. Liu, J. Lin, J.Y. Zhu, and S. Han, “Differentiable augmentation for data-efficient gan training,” Advances in Neural Information Processing Systems, vol.33, pp.7559–7570, 2020.
  4. T. Karras, M. Aittala, J. Hellsten, S. Laine, J. Lehtinen, and T. Aila, “Training generative adversarial networks with limited data,” Advances in Neural Information Processing Systems, vol.33, pp.12104– 12114, 2020.
  5. A. Razavi, A. Van den Oord, and O. Vinyals, “Generating diverse high-fidelity images with vq-vae-2,” Advances in neural information processing systems, vol.32, 2019.
  6. I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” Advances in neural information processing systems, vol.27, 2014.
  7. A. Aggarwal, M. Mittal, and G. Battineni, “Generative adversarial network: An overview of theory and applications,” International Journal of Information Management Data Insights, vol.1, no.1, p.100004, 2021.
  8. J. Sohl-Dickstein, E. Weiss, N. Maheswaranathan, and S. Ganguli, “Deep unsupervised learning using nonequilibrium thermodynamics,” International conference on machine learning, pp.2256–2265, PMLR, 2015.
  9. X. Chen, N. Mishra, M. Rohaninejad, and P. Abbeel, “Pixelsnail: An improved autoregressive generative model,” International Conference on Machine Learning, pp.864–872, PMLR, 2018.
  10. T. Karras, S. Laine, and T. Aila, “A style-based generator architecture for generative adversarial networks,” Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.4401– 4410, 2019.
  11. Y. Choi, Y. Uh, J. Yoo, and J.W. Ha, “Stargan v2: Diverse image synthesis for multiple domains,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2020.
  12. G. Parmar, R. Zhang, and J.Y. Zhu, “On aliased resizing and surprising subtleties in gan evaluation,” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.11410– 11420, 2022.
  13. T. Karras, M. Aittala, S. Laine, E. Härkönen, J. Hellsten, J. Lehtinen, and T. Aila, “Alias-free generative adversarial networks,” Proc. NeurIPS, 2021.
  14. A. Adadi, “A survey on data-efficient algorithms in big data era,” Journal of Big Data, vol.8, no.1, pp.1–54, 2021.
  15. C. Shorten and T.M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” Journal of big data, vol.6, no.1, pp.1–48, 2019.
  16. A. Mikołajczyk and M. Grochowski, “Data augmentation for improving deep learning in image classification problem,” 2018 international interdisciplinary PhD workshop (IIPhDW), pp.117–122, IEEE, 2018.
  17. E.D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q.V. Le, “Autoaugment: Learning augmentation policies from data,” arXiv preprint arXiv:1805.09501, 2018.
  18. H. Jun, R. Child, M. Chen, J. Schulman, A. Ramesh, A. Radford, and I. Sutskever, “Distribution augmentation for generative modeling,” International Conference on Machine Learning, pp.5006–5019, PMLR, 2020.
  19. OpenAI, “ChatGPT (September 25 Version).” https://chat. openai.com/chat, 2023. [Large language model].

Summary

We haven't generated a summary for this paper yet.