Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Are Natural Domain Foundation Models Useful for Medical Image Classification? (2310.19522v2)

Published 30 Oct 2023 in cs.CV

Abstract: The deep learning field is converging towards the use of general foundation models that can be easily adapted for diverse tasks. While this paradigm shift has become common practice within the field of natural language processing, progress has been slower in computer vision. In this paper we attempt to address this issue by investigating the transferability of various state-of-the-art foundation models to medical image classification tasks. Specifically, we evaluate the performance of five foundation models, namely SAM, SEEM, DINOv2, BLIP, and OpenCLIP across four well-established medical imaging datasets. We explore different training settings to fully harness the potential of these models. Our study shows mixed results. DINOv2 consistently outperforms the standard practice of ImageNet pretraining. However, other foundation models failed to consistently beat this established baseline indicating limitations in their transferability to medical image classification tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Joana Palés Huix (2 papers)
  2. Adithya Raju Ganeshan (1 paper)
  3. Johan Fredin Haslum (8 papers)
  4. Magnus Söderberg (4 papers)
  5. Christos Matsoukas (13 papers)
  6. Kevin Smith (43 papers)
Citations (14)