Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Comprehensive Survey on Test-Time Adaptation under Distribution Shifts (2303.15361v2)

Published 27 Mar 2023 in cs.LG, cs.AI, and cs.CV

Abstract: Machine learning methods strive to acquire a robust model during the training process that can effectively generalize to test samples, even in the presence of distribution shifts. However, these methods often suffer from performance degradation due to unknown test distributions. Test-time adaptation (TTA), an emerging paradigm, has the potential to adapt a pre-trained model to unlabeled data during testing, before making predictions. Recent progress in this paradigm has highlighted the significant benefits of using unlabeled data to train self-adapted models prior to inference. In this survey, we categorize TTA into several distinct groups based on the form of test data, namely, test-time domain adaptation, test-time batch adaptation, and online test-time adaptation. For each category, we provide a comprehensive taxonomy of advanced algorithms and discuss various learning scenarios. Furthermore, we analyze relevant applications of TTA and discuss open challenges and promising areas for future research. For a comprehensive list of TTA methods, kindly refer to \url{https://github.com/tim-learn/awesome-test-time-adaptation}.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Jian Liang (162 papers)
  2. Ran He (172 papers)
  3. Tieniu Tan (119 papers)
Citations (142)

Summary

Overview of "A Comprehensive Survey on Test-Time Adaptation under Distribution Shifts"

The paper presented by Liang, He, and Tan examines the burgeoning area of Test-Time Adaptation (TTA), a concept integral to coping with distribution shifts that challenge machine learning models' ability to generalize. This comprehensive survey classifies TTA methodologies into distinct paradigms, addresses their theoretical underpinnings, and analyzes various algorithms' practical applications.

Key Concepts and Definitions

Test-Time Adaptation seeks to modify a pre-trained model's capabilities to predict accurately on test data that differ in distribution from training data. This survey meticulously details TTA into several categories:

  1. Source-Free Domain Adaptation (SFDA): Involves leveraging a pre-trained model without access to source data, focusing on inferring labels for all samples in a target domain.
  2. Test-Time Batch Adaptation (TTBA): Adapts models to batches or individual instances at test time, capitalizing on similarities within a batch.
  3. Online Test-Time Adaptation (OTTA): Handles streaming data inputs in a sequential manner, with models adapting continuously.
  4. Test-Time Prior Adaptation (TTPA): Focuses on label distribution changes, aiming to correct posteriors based on estimated label priors.

Taxonomy of Algorithms

The survey categorizes SFDA methods into pseudo-labeling techniques, consistency regularization, clustering-based training, source distribution estimation, and self-supervised learning. Each method aims to bypass the need for source data by employing diverse techniques like entropy minimization, mutual information maximization, and neighborhood clustering.

TTBA methods are critically examined concerning their strategies such as batch normalization calibration, model optimization through auxiliary tasks, and input adaptation. These techniques are foundational in dynamically adjusting models during test phases.

OTTA approaches are defined by their capability to adapt under both stationary and dynamic distribution shifts. The survey highlights batch normalization adjustments and entropy minimization while spotlighting anti-forgetting regularization to negate the risk of catastrophic forgetting.

TTPA strategies revolve around prior estimation techniques, including confusion matrix applications and maximum likelihood estimations, ensuring the adaptation to label distribution drifts.

Practical Applications

The surveyed methodologies find their applications across various domains:

  • Image Classification and Segmentation: Utilizing benchmark datasets under different conditions to validate proposed TTA techniques.
  • Video, Medical, and Multi-Modal Data: Extending TTA applications to real-time processing, 3D segmentation, and multi-modal analyses.
  • Low-Level Vision Tasks and Generative Models: Integration of TTA in tasks like super-resolution and style transfer.

Challenges and Future Directions

The survey identifies critical challenges within TTA, such as the need for standard benchmarks and validation protocols, the potential extensions to big models, and adaptation goals like fairness and privacy. Future research is anticipated to explore these aspects, paving the way for more robust and versatile TTA techniques.

Conclusion

This paper offers a detailed exploration into TTA, presenting a clear taxonomy, identifying challenges, and suggesting future research directions. It stands as a pivotal reference for researchers exploring model adaptability in dynamically changing environments and distribution shifts. By providing a thorough overview of the domain, it lays the groundwork for future advancements in the development of adaptive and versatile AI systems.