Boosting Meta-Training with Base Class Information for Few-Shot Learning (2403.03472v1)
Abstract: Few-shot learning, a challenging task in machine learning, aims to learn a classifier adaptable to recognize new, unseen classes with limited labeled examples. Meta-learning has emerged as a prominent framework for few-shot learning. Its training framework is originally a task-level learning method, such as Model-Agnostic Meta-Learning (MAML) and Prototypical Networks. And a recently proposed training paradigm called Meta-Baseline, which consists of sequential pre-training and meta-training stages, gains state-of-the-art performance. However, as a non-end-to-end training method, indicating the meta-training stage can only begin after the completion of pre-training, Meta-Baseline suffers from higher training cost and suboptimal performance due to the inherent conflicts of the two training stages. To address these limitations, we propose an end-to-end training paradigm consisting of two alternative loops. In the outer loop, we calculate cross entropy loss on the entire training set while updating only the final linear layer. In the inner loop, we employ the original meta-learning training mode to calculate the loss and incorporate gradients from the outer loss to guide the parameter updates. This training paradigm not only converges quickly but also outperforms existing baselines, indicating that information from the overall training set and the meta-learning training paradigm could mutually reinforce one another. Moreover, being model-agnostic, our framework achieves significant performance gains, surpassing the baseline systems by approximate 1%.
- G. Koch, R. Zemel, R. Salakhutdinov et al., “Siamese neural networks for one-shot image recognition,” in ICML Deep Learning Workshop, 2015.
- B. M. Lake, R. Salakhutdinov, J. Gross, and J. B. Tenenbaum, “One shot learning of simple visual concepts,” in Proceedings of the 33th Annual Meeting of the Cognitive Science Society, 2011.
- F. Sung, Y. Yang, L. Zhang, T. Xiang, P. H. S. Torr, and T. M. Hospedales, “Learning to compare: Relation network for few-shot learning,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018.
- P. Tian, W. Li, and Y. Gao, “Consistent meta-regularization for better meta-knowledge in few-shot learning,” IEEE Transactions on Neural Networks and Learning Systems, vol. 33, no. 12, pp. 7277–7288, 2022.
- R.-Q. Wang, X.-Y. Zhang, and C.-L. Liu, “Meta-prototypical learning for domain-agnostic few-shot recognition,” IEEE Transactions on Neural Networks and Learning Systems, vol. 33, no. 11, pp. 6990–6996, 2022.
- J. Snell, K. Swersky, and R. S. Zemel, “Prototypical networks for few-shot learning,” in Advances in Neural Information Processing Systems, 2017.
- Y. Chen, Z. Liu, H. Xu, T. Darrell, and X. Wang, “Meta-baseline: Exploring simple meta-learning for few-shot learning,” in IEEE/CVF International Conference on Computer Vision, 2021.
- R. Hou, H. Chang, B. Ma, S. Shan, and X. Chen, “Cross attention network for few-shot classification,” in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019, pp. 4005–4016.
- H. Ye, H. Hu, D. Zhan, and F. Sha, “Few-shot learning via embedding adaptation with set-to-set functions,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
- C. Finn, P. Abbeel, and S. Levine, “Model-agnostic meta-learning for fast adaptation of deep networks,” in Proceedings of the 34th International Conference on Machine Learning, 2017.
- Z. Li, F. Zhou, F. Chen, and H. Li, “Meta-sgd: Learning to learn quickly for few shot learning,” CoRR, vol. abs/1707.09835, 2017. [Online]. Available: http://arxiv.org/abs/1707.09835
- A. Nichol, J. Achiam, and J. Schulman, “On first-order meta-learning algorithms,” CoRR, vol. abs/1803.02999, 2018. [Online]. Available: http://arxiv.org/abs/1803.02999
- S. Ravi and H. Larochelle, “Optimization as a model for few-shot learning,” in 5th International Conference on Learning Representations, 2017.
- A. Raghu, M. Raghu, S. Bengio, and O. Vinyals, “Rapid learning or feature reuse? towards understanding the effectiveness of MAML,” in 8th International Conference on Learning Representations, 2020.
- J. Oh, H. Yoo, C. Kim, and S. Yun, “BOIL: towards representation change for few-shot learning,” in 9th International Conference on Learning Representations, 2021.
- A. A. Rusu, D. Rao, J. Sygnowski, O. Vinyals, R. Pascanu, S. Osindero, and R. Hadsell, “Meta-learning with latent embedding optimization,” in 7th International Conference on Learning Representations, 2019.
- W. Li, L. Wang, J. Xu, J. Huo, Y. Gao, and J. Luo, “Revisiting local descriptor based image-to-class measure for few-shot learning,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019.
- B. N. Oreshkin, P. R. López, and A. Lacoste, “TADAM: task dependent adaptive metric for improved few-shot learning,” in Advances in Neural Information Processing Systems, 2018.
- O. Vinyals, C. Blundell, T. Lillicrap, k. kavukcuoglu, and D. Wierstra, “Matching networks for one shot learning,” in Advances in Neural Information Processing Systems, 2016.
- Y. Liu, J. Lee, M. Park, S. Kim, E. Yang, S. J. Hwang, and Y. Yang, “Learning to propagate labels: Transductive propagation network for few-shot learning,” in 7th International Conference on Learning Representations, 2019.
- C. Zhang, Y. Cai, G. Lin, and C. Shen, “Deepemd: Few-shot image classification with differentiable earth mover’s distance and structured classifiers,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
- S. Gidaris and N. Komodakis, “Dynamic few-shot visual learning without forgetting,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018.
- W. Chen, Y. Liu, Z. Kira, Y. F. Wang, and J. Huang, “A closer look at few-shot classification,” in 7th International Conference on Learning Representations, 2019.
- Y. Tian, Y. Wang, D. Krishnan, J. B. Tenenbaum, and P. Isola, “Rethinking few-shot image classification: a good embedding is all you need?” in Computer Vision–ECCV 2020: 16th European Conference, 2020.
- G. S. Dhillon, P. Chaudhari, A. Ravichandran, and S. Soatto, “A baseline for few-shot image classification,” in 8th International Conference on Learning Representations, 2020.
- Z. Shen, Z. Liu, J. Qin, M. Savvides, and K. Cheng, “Partial is better than all: Revisiting fine-tuning strategy for few-shot learning,” in Thirty-Fifth AAAI Conference on Artificial Intelligence, 2021.
- N. Lai, M. Kan, C. Han, X. Song, and S. Shan, “Learning to learn adaptive classifier–predictor for few-shot learning,” IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 8, pp. 3458–3470, 2021.
- J. Xie, F. Long, J. Lv, Q. Wang, and P. Li, “Joint distribution matters: Deep brownian distance covariance for few-shot classification,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022.
- S. X. Hu, D. Li, J. Stühmer, M. Kim, and T. M. Hospedales, “Pushing the limits of simple pipelines for few-shot learning: External data and fine-tuning make a difference,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022.
- Z. Yang, J. Wang, and Y. Zhu, “Few-shot classification with contrastive learning,” in Computer Vision - ECCV 2022 - 17th European Conference, 2022.
- D. Wertheimer, L. Tang, and B. Hariharan, “Few-shot classification with feature map reconstruction networks,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
- C. Wah, S. Branson, P. Welinder, P. Perona, and S. Belongie, “The caltech-ucsd birds-200-2011 dataset,” 2011.
- M. Ren, E. Triantafillou, S. Ravi, J. Snell, K. Swersky, J. B. Tenenbaum, H. Larochelle, and R. S. Zemel, “Meta-learning for semi-supervised few-shot classification,” in 6th International Conference on Learning Representations, 2018.
- A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems, 2012.
- K. Lee, S. Maji, A. Ravichandran, and S. Soatto, “Meta-learning with differentiable convex optimization,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019.
- A. Ravichandran, R. Bhotika, and S. Soatto, “Few-shot learning with embedded class models and shot-free meta training,” in IEEE/CVF International Conference on Computer Vision, 2019.
- Q. Sun, Y. Liu, T. Chua, and B. Schiele, “Meta-transfer learning for few-shot learning,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019.
- H. Lee, S. J. Hwang, and J. Shin, “Self-supervised label augmentation via input transformations,” in Proceedings of the 37th International Conference on Machine Learning, 2020.
- A. Li, W. Huang, X. Lan, J. Feng, Z. Li, and L. Wang, “Boosting few-shot learning with adaptive margin loss,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
- C. Xing, N. Rostamzadeh, B. N. Oreshkin, and P. O. Pinheiro, “Adaptive cross-modal few-shot learning,” in Advances in Neural Information Processing Systems, 2019.
- C. Simon, P. Koniusz, R. Nock, and M. Harandi, “Adaptive subspaces for few-shot learning,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
- J. Su, S. Maji, and B. Hariharan, “When does self-supervision improve few-shot learning?” in Computer Vision - ECCV 2020 - 16th European Conference, 2020.
- Weihao Jiang (12 papers)
- Guodong Liu (133 papers)
- Di He (108 papers)
- Kun He (177 papers)