Flatten Long-Range Loss Landscapes for Cross-Domain Few-Shot Learning (2403.00567v2)
Abstract: Cross-domain few-shot learning (CDFSL) aims to acquire knowledge from limited training data in the target domain by leveraging prior knowledge transferred from source domains with abundant training samples. CDFSL faces challenges in transferring knowledge across dissimilar domains and fine-tuning models with limited training data. To address these challenges, we initially extend the analysis of loss landscapes from the parameter space to the representation space, which allows us to simultaneously interpret the transferring and fine-tuning difficulties of CDFSL models. We observe that sharp minima in the loss landscapes of the representation space result in representations that are hard to transfer and fine-tune. Moreover, existing flatness-based methods have limited generalization ability due to their short-range flatness. To enhance the transferability and facilitate fine-tuning, we introduce a simple yet effective approach to achieve long-range flattening of the minima in the loss landscape. This approach considers representations that are differently normalized as minima in the loss landscape and flattens the high-loss region in the middle by randomly sampling interpolated representations. We implement this method as a new normalization layer that replaces the original one in both CNNs and ViTs. This layer is simple and lightweight, introducing only a minimal number of additional parameters. Experimental results on 8 datasets demonstrate that our approach outperforms state-of-the-art methods in terms of average accuracy. Moreover, our method achieves performance improvements of up to 9\% compared to the current best approaches on individual datasets. Our code will be released.
- Adversarial attacks and defences: A survey, 2018.
- A closer look at few-shot classification. In Proceedings of the International Conference on Learning Representations, 2018.
- On separate normalization in self-supervised transformers, 2023.
- Meta batch-instance normalization for generalizable person re-identification, 2021.
- Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic), 2019.
- Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 248–255. Ieee, 2009.
- An image is worth 16x16 words: Transformers for image recognition at scale, 2021.
- Sharpness-aware minimization for efficiently improving generalization, 2021.
- Wave-san: Wavelet based style augmentation network for cross-domain few-shot learning, 2022.
- Styleadv: Meta style adversarial training for cross-domain few-shot learning, 2023.
- A broader study of cross-domain few-shot learning. In Proceedings of the IEEE/CVF European Conference on Computer Vision, pages 124–141. Springer, 2020.
- Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 770–778, 2016.
- Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification, 2019.
- Benchmarking neural network robustness to common corruptions and perturbations, 2019.
- The inaturalist species classification and detection dataset, 2018.
- Adversarial feature augmentation for cross-domain few-shot classification, 2022.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, pages 448–456. PMLR, 2015.
- On large-batch training for deep learning: Generalization gap and sharp minima, 2017.
- 3d object representations for fine-grained categorization. In 2013 IEEE International Conference on Computer Vision Workshops, pages 554–561, 2013.
- Ranking distance calibration for cross-domain few-shot learning, 2022.
- Decoupled weight decay regularization, 2019.
- Using deep learning for image-based plant disease detection. Frontiers in Plant Science, 7(September), 2016. Publisher Copyright: © 2016 Mohanty, Hughes and Salathé.
- Jan Stühmer Shell Xu, Da Li. Pushing the limits of simple pipelines for few-shot learning: External data and fine-tuning make a difference, 2022.
- Overcoming catastrophic forgetting in incremental few-shot learning by finding flat minima, 2021.
- Cross-domain few-shot classification via learned feature-wise transformation. In Proceedings of the International Conference on Learning Representations, 2020.
- Instance normalization: The missing ingredient for fast stylization, 2017.
- Manifold mixup: Better representations by interpolating hidden states, 2019.
- Matching networks for one shot learning. In Proceedings of the International Conference on Neural Information Processing Systems, pages 3637–3645, 2016.
- The caltech-ucsd birds-200-2011 dataset. 2011.
- Cross-domain few-shot classification via adversarial task augmentation, 2021.
- ChestX-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2017.
- mixup: Beyond empirical risk minimization, 2018.
- Dino: Detr with improved denoising anchor boxes for end-to-end object detection, 2022.
- Gradient norm aware minimization seeks first-order flatness and improves generalization, 2023.
- Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(6):1452–1464, 2018.
- Revisiting prototypical network for cross domain few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 20061–20070, 2023.
- Margin-based few-shot class-incremental learning with class-level overfitting mitigation. In Proceedings of the International Conference on Neural Information Processing Systems, 2022.
- Yixiong Zou (28 papers)
- Yicong Liu (4 papers)
- Yiman Hu (1 paper)
- Yuhua Li (29 papers)
- Ruixuan Li (60 papers)