Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Covariance-aware Feature Alignment with Pre-computed Source Statistics for Test-time Adaptation to Multiple Image Corruptions (2204.13263v2)

Published 28 Apr 2022 in cs.LG

Abstract: Real-world image recognition systems often face corrupted input images, which cause distribution shifts and degrade the performance of models. These systems often use a single prediction model in a central server and process images sent from various environments, such as cameras distributed in cities or cars. Such single models face images corrupted in heterogeneous ways in test time. Thus, they require to instantly adapt to the multiple corruptions during testing rather than being re-trained at a high cost. Test-time adaptation (TTA), which aims to adapt models without accessing the training dataset, is one of the settings that can address this problem. Existing TTA methods indeed work well on a single corruption. However, the adaptation ability is limited when multiple types of corruption occur, which is more realistic. We hypothesize this is because the distribution shift is more complicated, and the adaptation becomes more difficult in case of multiple corruptions. In fact, we experimentally found that a larger distribution gap remains after TTA. To address the distribution gap during testing, we propose a novel TTA method named Covariance-Aware Feature alignment (CAFe). We empirically show that CAFe outperforms prior TTA methods on image corruptions, including multiple types of corruptions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. “Benchmarking Neural Network Robustness to Common Corruptions and Perturbations,” in Proceedings of the International Conference on Learning Representations (ICLR), 2019.
  2. “Do imagenet classifiers generalize to imagenet?,” in International Conference on Machine Learning (ICML), 2019.
  3. G Sreenu and Saleem Durai, “Intelligent video surveillance: a review through deep learning techniques for crowd analysis,” Journal of Big Data, vol. 6, no. 1, pp. 1–27, 2019.
  4. “Improving robustness against common corruptions by covariate shift adaptation,” in Advances in Neural Information Processing Systems, 2020.
  5. “Revisiting Batch Normalization For Practical Domain Adaptation,” in International Conference on Learning Representations Workshop, 2017.
  6. “Revisiting Batch Normalization for Improving Corruption Robustness,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2021.
  7. “Tent: Fully Test-Time Adaptation by Entropy Minimization,” in Proceedings of the International Conference on Learning Representations (ICLR), 2021.
  8. “Bayesian Adaptation for Covariate Shift,” in Advances in Neural Information Processing Systems, 2021.
  9. “Test-time classifier adjustment module for model-agnostic domain generalization,” in Advances in Neural Information Processing Systems, 2021.
  10. “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift,” in Proceedings of the International Conference on Machine Learning (ICML), 2015.
  11. “A theory of learning from different domains,” Machine learning, vol. 79, no. 1, pp. 151–175, 2010.
  12. “Spectral partitioning: The more eigenvectors, the better,” in Proceedings of the 32nd annual ACM/IEEE design automation conference, 1995.
  13. Gabriela Csurka, “Domain adaptation for visual applications: A comprehensive survey,” arXiv preprint arXiv:1702.05374, 2017.
  14. “Domain-adversarial training of neural networks,” The Journal of Machine Learning Research, vol. 17, no. 1, pp. 2096–2030, 2016.
  15. “Deep coral: Correlation alignment for deep domain adaptation,” in European conference on computer vision. Springer, 2016, pp. 443–450.
  16. “Return of frustratingly easy domain adaptation,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2016, vol. 30.
  17. “KL Guided Domain Adaptation,” in Proceedings of the International Conference on Learning Representations (ICLR), 2022.
  18. “Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation,” in Proceedings of the International Conference on Machine Learning (ICML), 2020.
  19. “Model Adaptation: Unsupervised Domain Adaptation Without Source Data,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
  20. “Universal source-free domain adaptation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
  21. “Source-Free Adaptation to Measurement Shift via Bottom-Up Feature Restoration,” in Proceedings of the International Conference on Learning Representations (ICLR), 2022.
  22. John Duchi, “Derivations for Linear Algebra and Optimization,” 2007, http://ai.stanford.edu/~jduchi/projects/general_notes.pdf (visited on June 22nd, 2023).
  23. “Unsupervised Classifiers, Mutual Information and 'Phantom Targets,” in Advances in Neural Information Processing Systems, 1992.
  24. “Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
  25. “Semi-supervised domain adaptation via minimax entropy,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019.
  26. “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 2016.
  27. “ImageNet Large Scale Visual Recognition Challenge,” International Journal of Computer Vision (IJCV), vol. 115, no. 3, pp. 211–252, 2015.
  28. “Scikit-learn: Machine Learning in Python,” Journal of Machine Learning Research, vol. 12, no. 85, pp. 2825–2830, 2011.
  29. “PyTorch: An Imperative Style, High-Performance Deep Learning Library,” in Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, Eds., pp. 8024–8035. Curran Associates, Inc., 2019.
  30. “High-level library to help with training neural networks in pytorch,” https://github.com/pytorch/ignite, 2020.
  31. Ross Wightman, “Pytorch image models,” https://github.com/rwightman/pytorch-image-models, 2019.
  32. TorchVision maintainers and contributors, “Torchvision: Pytorch’s computer vision library,” https://github.com/pytorch/vision, 2016.
  33. Frank Hutter Ilya Loshchilov, “SGDR: stochastic gradient descent with warm restarts,” in International Conference on Learning Representations (ICLR), 2017.
  34. “A simple baseline for bayesian uncertainty in deep learning,” Advances in Neural Information Processing Systems, vol. 32, pp. 13153–13164, 2019.
  35. “The Fréchet distance between multivariate normal distributions,” Journal of Multivariate Analysis, vol. 12, no. 3, pp. 450–455, 1982.
  36. “Learning multiple layers of features from tiny images,” Master’s thesis, University of Tront, 2009.
  37. “Reading digits in natural images with unsupervised feature learning,” 2011.
  38. “MNIST handwritten digit database,” 1998, http://yann.lecun.com/exdb/mnist/.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Kazuki Adachi (10 papers)
  2. Shin'ya Yamaguchi (24 papers)
  3. Atsutoshi Kumagai (22 papers)