Robust and Efficient Medical Imaging with Self-Supervision (2205.09723v2)
Abstract: Recent progress in Medical AI has delivered systems that can reach clinical expert level performance. However, such systems tend to demonstrate sub-optimal "out-of-distribution" performance when evaluated in clinical settings different from the training environment. A common mitigation strategy is to develop separate systems for each clinical setting using site-specific data [1]. However, this quickly becomes impractical as medical data is time-consuming to acquire and expensive to annotate [2]. Thus, the problem of "data-efficient generalization" presents an ongoing difficulty for Medical AI development. Although progress in representation learning shows promise, their benefits have not been rigorously studied, specifically for out-of-distribution settings. To meet these challenges, we present REMEDIS, a unified representation learning strategy to improve robustness and data-efficiency of medical imaging AI. REMEDIS uses a generic combination of large-scale supervised transfer learning with self-supervised learning and requires little task-specific customization. We study a diverse range of medical imaging tasks and simulate three realistic application scenarios using retrospective data. REMEDIS exhibits significantly improved in-distribution performance with up to 11.5% relative improvement in diagnostic accuracy over a strong supervised baseline. More importantly, our strategy leads to strong data-efficient generalization of medical imaging AI, matching strong supervised baselines using between 1% to 33% of retraining data across tasks. These results suggest that REMEDIS can significantly accelerate the life-cycle of medical imaging AI development thereby presenting an important step forward for medical imaging AI to deliver broad impact.
- Shekoofeh Azizi (23 papers)
- Laura Culp (8 papers)
- Jan Freyberg (14 papers)
- Basil Mustafa (32 papers)
- Sebastien Baur (7 papers)
- Simon Kornblith (53 papers)
- Ting Chen (148 papers)
- Patricia MacWilliams (4 papers)
- Ellery Wulczyn (14 papers)
- Boris Babenko (9 papers)
- Megan Wilson (3 papers)
- Aaron Loh (5 papers)
- Po-Hsuan Cameron Chen (10 papers)
- Yuan Liu (342 papers)
- Pinal Bavishi (4 papers)
- Scott Mayer McKinney (8 papers)
- Jim Winkens (6 papers)
- Abhijit Guha Roy (28 papers)
- Zach Beaver (2 papers)
- Fiona Ryan (13 papers)