AdaAugment: A Tuning-Free and Adaptive Approach to Enhance Data Augmentation (2405.11467v2)
Abstract: Data augmentation (DA) is widely employed to improve the generalization performance of deep models. However, most existing DA methods use augmentation operations with random magnitudes throughout training. While this fosters diversity, it can also inevitably introduce uncontrolled variability in augmented data, which may cause misalignment with the evolving training status of the target models. Both theoretical and empirical findings suggest that this misalignment increases the risks of underfitting and overfitting. To address these limitations, we propose AdaAugment, an innovative and tuning-free Adaptive Augmentation method that utilizes reinforcement learning to dynamically adjust augmentation magnitudes for individual training samples based on real-time feedback from the target network. Specifically, AdaAugment features a dual-model architecture consisting of a policy network and a target network, which are jointly optimized to effectively adapt augmentation magnitudes. The policy network optimizes the variability within the augmented data, while the target network utilizes the adaptively augmented samples for training. Extensive experiments across benchmark datasets and deep architectures demonstrate that AdaAugment consistently outperforms other state-of-the-art DA methods in effectiveness while maintaining remarkable efficiency.
- Clutr: Curriculum learning via unsupervised task representation learning. In International Conference on Machine Learning, pp. 1361–1395. PMLR, 2023.
- An information-theoretic perspective on overfitting and underfitting. In Advances in Artificial Intelligence: 33rd Australasian Joint Conference, pp. 347–358. Springer, 2020.
- Byeon, H. Advances in value-based, policy-based, and deep learning-based reinforcement learning. International Journal of Advanced Computer Science and Applications, 14(8), 2023.
- Gridmask data augmentation. arXiv preprint arXiv:2001.04086, 2020.
- Adaaug: Learning class-and instance-adaptive data augmentation policies. In International Conference on Learning Representations, 2021.
- A downsampled variant of imagenet as an alternative to the cifar datasets. arXiv preprint arXiv:1707.08819, Aug 2017.
- Autoaugment: Learning augmentation strategies from data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 113–123, 2019.
- Randaugment: Practical automated data augmentation with a reduced search space. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M. F., and Lin, H. (eds.), Proc. Adv. Neural Inf. Process. Syst., volume 33, pp. 18613–18624, 2020.
- Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552, 2017.
- Adaptive data augmentation for image classification. In 2016 IEEE international conference on image processing (ICIP), pp. 3688–3692. Ieee, 2016.
- Gastaldi, X. Shake-shake regularization. 2017.
- Keepaugment: A simple information-preserving data augmentation approach. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1055–1064, 2021.
- A survey of actor-critic reinforcement learning: Standard and natural policy gradients. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 42(6):1291–1307, 2012.
- Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 770–778, 2016.
- When to learn what: Model-adaptive data augmentation curriculum. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1717–1728, 2023.
- Feature map testing for deep neural networks. arXiv preprint arXiv:2307.11563, 2023.
- Deep reinforcement learning for autonomous driving: A survey. IEEE Transactions on Intelligent Transportation Systems, 23(6):4909–4926, 2022. doi: 10.1109/TITS.2021.3054625.
- Collecting a large-scale dataset of fine-grained cars. 2013.
- Convolutional deep belief networks on cifar-10. Unpublished manuscript, 40(7):1–9, 2010.
- Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6):84–90, 2017.
- Deep reinforcement learning in computer vision: a comprehensive survey. Artificial Intelligence Review, pp. 1–87, 2022.
- The MNIST database of handwritten digits, 1998. URL http://yann. lecun. com/exdb/mnist, 10(34):14, 1998.
- Learning augmentation network via influence functions. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10961–10970, 2020.
- Keeping deep learning models in check: A history-based approach to mitigate overfitting, 2024.
- Li, S. E. Deep reinforcement learning. In Reinforcement Learning for Sequential Decision and Optimal Control, pp. 365–402. Springer, 2023.
- Fast autoaugment. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R. (eds.), Advances in neural information processing systems, volume 32. Curran Associates, Inc., 2019.
- Selectaugment: Hierarchical deterministic sample selection for data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pp. 1604–1612, 2023a.
- A survey on reinforcement learning for recommender systems. IEEE Transactions on Neural Networks and Learning Systems, 2023b.
- Data-efficient augmentation for training neural networks. Advances in Neural Information Processing Systems, 35:5124–5136, 2022.
- Large-scale long-tailed recognition in an open world. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- Fine-grained visual classification of aircraft. arXiv preprint arXiv:1306.5151, 2013.
- Asynchronous methods for deep reinforcement learning. In International conference on machine learning, pp. 1928–1937. PMLR, 2016.
- Trivialaugment: Tuning-free yet state-of-the-art data augmentation. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 774–782, 2021.
- Parallel and scalable dunn index for the validation of big data clusters. Parallel Computing, 102:102751, 2021.
- Automated flower classification over a large number of classes. In 2008 Sixth Indian conference on computer vision, graphics & image processing, pp. 722–729. IEEE, 2008.
- Cats and dogs. In 2012 IEEE conference on computer vision and pattern recognition, pp. 3498–3505. IEEE, 2012.
- Meta-learning requires meta-augmentation. Advances in Neural Information Processing Systems, 33:5705–5715, 2020.
- Undecidability of underfitting in learning algorithms. In 2021 2nd International Conference on Computing and Data Science (CDS), pp. 591–594. IEEE, 2021.
- Reinforcement learning algorithms: A brief survey. Expert Systems with Applications, pp. 120495, 2023.
- Hide-and-seek: Forcing a network to be meticulous for weakly-supervised object and action localization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3544–3553. IEEE, Oct 2017.
- Curriculum learning: A survey. International Journal of Computer Vision, 130(6):1526–1565, 2022.
- Policy gradient methods for reinforcement learning with function approximation. Advances in neural information processing systems, 12, 1999.
- Suzuki, T. Teachaugment: Data augmentation optimization using teacher knowledge. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10904–10914, 2022.
- Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1492–1500, 2017.
- A comprehensive survey of image augmentation techniques for deep learning. Pattern Recognition, 137:109347, 2023. ISSN 0031-3203. doi: https://doi.org/10.1016/j.patcog.2023.109347.
- Universal adaptive data augmentation. arXiv preprint arXiv:2207.06658, 2022.
- Image data augmentation for deep learning: A survey. arXiv preprint arXiv:2204.08610, 2022.
- Advmask: A sparse adversarial attack-based data augmentation method for image classification. Pattern Recognition, 144:109847, 2023.
- Investigating the effectiveness of data augmentation from similarity and diversity: An empirical study. Pattern Recognition, 148:110204, 2024. ISSN 0031-3203. doi: https://doi.org/10.1016/j.patcog.2023.110204.
- Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 6023–6032, 2019.
- Wide residual networks. In Richard C. Wilson, E. R. H. and Smith, W. A. P. (eds.), Proc. Brit. Mach. Vis. Conf. (BMVC), pp. 87.1–87.12. BMVA Press, September 2016. ISBN 1-901725-59-6. doi: 10.5244/C.30.87. URL https://dx.doi.org/10.5244/C.30.87.
- A-fmi: learning attributions from deep networks via feature map importance. arXiv preprint arXiv:2104.05527, 2021.
- mixup: Beyond empirical risk minimization. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=r1Ddp1-Rb.
- Adversarial autoaugment. arXiv preprint arXiv:1912.11188, 2019.
- Adaptive data augmentation for contrastive learning. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1–5. IEEE, 2023.
- Random erasing data augmentation. In Proc. AAAI, volume 34, pp. 13001–13008, 2020.
- A comprehensive survey on transfer learning. Proc. IEEE, 109(1):43–76, 2021. doi: 10.1109/JPROC.2020.3004555.