Feature Alignment: Rethinking Efficient Active Learning via Proxy in the Context of Pre-trained Models (2403.01101v2)
Abstract: Fine-tuning the pre-trained model with active learning holds promise for reducing annotation costs. However, this combination introduces significant computational costs, particularly with the growing scale of pre-trained models. Recent research has proposed proxy-based active learning, which pre-computes features to reduce computational costs. Yet, this approach often incurs a significant loss in active learning performance, sometimes outweighing the computational cost savings. This paper demonstrates that not all sample selection differences result in performance degradation. Furthermore, we show that suitable training methods can mitigate the decline of active learning performance caused by certain selection discrepancies. Building upon detailed analysis, we propose a novel method, aligned selection via proxy, which improves proxy-based active learning performance by updating pre-computed features and selecting a proper training method. Extensive experiments validate that our method improves the total cost of efficient active learning while maintaining computational efficiency. The code is available at \url{https://github.com/ZiTingW/asvp}.
- Deep batch active learning by diverse, uncertain gradient lower bounds. In International Conference on Learning Representations, 2020.
- Reducing label effort: Self-supervised meets active learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1631–1639, 2021.
- A survey on active learning and human-in-the-loop deep learning for medical image analysis. Medical Image Analysis, 71:102062, 2021.
- Exponential moving average normalization for self-supervised and semi-supervised learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 194–203, 2021.
- Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pages 9650–9660, 2021.
- On the marginal benefit of active learning: Does self-supervision eat its cake? In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 3455–3459. IEEE, 2021.
- A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597–1607. PMLR, 2020a.
- Big self-supervised models are strong semi-supervised learners. Advances in neural information processing systems, 33:22243–22255, 2020b.
- Batch active learning at scale. Advances in Neural Information Processing Systems, 34:11933–11944, 2021.
- Selection via proxy: Efficient data selection for deep learning. In International Conference on Learning Representations, 2019.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- Active learning at the imagenet scale. arXiv preprint arXiv:2111.12880, 2021.
- Deep bayesian active learning with image data. In International Conference on Machine Learning, pages 1183–1192. PMLR, 2017.
- Bootstrap your own latent-a new approach to self-supervised learning. Advances in Neural Information Processing Systems, 33:21271–21284, 2020.
- Active learning on a budget: Opposite strategies suit high and low budgets. In International Conference on Machine Learning, pages 8175–8195. PMLR, 2022.
- Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000–16009, 2022.
- Batchbald: Efficient and diverse batch acquisition for deep bayesian active learning. Advances in neural information processing systems, 32, 2019.
- Learning multiple layers of features from tiny images. Technical Report TR-2009, University of Toronto, 2009.
- Fine-tuning can distort pretrained features and underperform out-of-distribution. In International Conference on Learning Representations, 2022.
- Heterogeneous uncertainty sampling for supervised learning. In Machine learning proceedings 1994, pages 148–156. Elsevier, 1994.
- Exploring model transferability through the lens of potential energy. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5429–5438, 2023.
- Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 10012–10022, 2021.
- Low-budget active learning via wasserstein distance: An integer programming approach. In International Conference on Learning Representations, 2022.
- A deep active learning system for species identification and counting in camera trap images. Methods in ecology and evolution, 12(1):150–161, 2021.
- Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193, 2023.
- Cats and dogs. In 2012 IEEE conference on computer vision and pattern recognition, pages 3498–3505. IEEE, 2012.
- A survey of deep active learning. ACM computing surveys (CSUR), 54(9):1–40, 2021.
- How to prepare your task head for finetuning. In The Eleventh International Conference on Learning Representations, 2022.
- Active hidden markov models for information extraction. In International symposium on intelligent data analysis, pages 309–318. Springer, 2001.
- Active learning for convolutional neural networks: A core-set approach. In International Conference on Learning Representations, 2018.
- Deep active learning: Unified and principled method for query and training. In International Conference on Artificial Intelligence and Statistics, pages 1308–1318. PMLR, 2020.
- Variational adversarial active learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5972–5981, 2019.
- Bayesian generative active deep learning. In International Conference on Machine Learning, pages 6295–6304. PMLR, 2019.
- Ntkcpl: Active learning on top of self-supervised model by estimating true coverage. arXiv preprint arXiv:2306.04099, 2023.
- Jae Oh Woo. Active learning in bayesian neural networks with balanced entropy learning principle. In The Eleventh International Conference on Learning Representations, 2022.
- Active finetuning: Exploiting annotation budget in the pretraining-finetuning paradigm. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 23715–23724, 2023.
- Suggestive annotation: A deep active learning framework for biomedical image segmentation. In International conference on medical image computing and computer-assisted intervention, pages 399–407. Springer, 2017.
- Active learning through a covering lens. In Advances in Neural Information Processing Systems, 2022.
- Learning loss for active learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 93–102, 2019.
- Logme: Practical assessment of pre-trained models for transfer learning. In International Conference on Machine Learning, pages 12133–12143. PMLR, 2021.
- Labelbench: A comprehensive framework for benchmarking label-efficient learning. arXiv preprint arXiv:2306.09910, 2023.
- Scalable batch-mode deep bayesian active learning via equivalence class annealing. In The Eleventh International Conference on Learning Representations, 2022.