Is in-domain data beneficial in transfer learning for landmarks detection in x-ray images? (2403.01470v1)
Abstract: In recent years, deep learning has emerged as a promising technique for medical image analysis. However, this application domain is likely to suffer from a limited availability of large public datasets and annotations. A common solution to these challenges in deep learning is the usage of a transfer learning framework, typically with a fine-tuning protocol, where a large-scale source dataset is used to pre-train a model, further fine-tuned on the target dataset. In this paper, we present a systematic study analyzing whether the usage of small-scale in-domain x-ray image datasets may provide any improvement for landmark detection over models pre-trained on large natural image datasets only. We focus on the multi-landmark localization task for three datasets, including chest, head, and hand x-ray images. Our results show that using in-domain source datasets brings marginal or no benefit with respect to an ImageNet out-of-domain pre-training. Our findings can provide an indication for the development of robust landmark detection systems in medical images when no large annotated dataset is available.
- “Deepnavnet: Automated landmark localization for neuronavigation,” Frontiers in Neuroscience, vol. 15, pp. 670287, 2021.
- “Two-stage mesh deep learning for automated tooth segmentation and landmark localization on 3d intraoral scans,” IEEE Trans. Medical Imaging, vol. 41, no. 11, pp. 3158–3166, 2022.
- “Cephalometric landmark detection via global and local encoders and patch-wise attentions,” Neurocomputing, vol. 470, pp. 182–189, 2022.
- “You only learn once: Universal anatomical landmark detection,” in International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, 2021, pp. 85–95.
- “A study of the effectiveness of transfer learning in individualized asthma risk prediction,” in Proceedings of the 36th Annual ACM Symposium on Applied Computing, New York, NY, USA, 2021, SAC ’21, p. 1082–1085, Association for Computing Machinery.
- “Top-tuning: A study on transfer learning for an efficient alternative to fine tuning for image classification with fast kernel methods,” Image and Vision Computing, vol. 142, pp. 104894, 2024.
- “Taskonomy: Disentangling task transfer learning,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3712–3722.
- “Food image classification: The benefit of in-domain transfer learning,” in International Conference on Image Analysis and Processing. Springer, 2023, pp. 259–269.
- “KNEEL: knee anatomical landmark localization using hourglass networks,” CoRR, vol. abs/1907.12237, 2019.
- “Deep learning approach for automatic landmark detection and alignment analysis in whole-spine lateral radiographs,” Scientific Reports, vol. 11, no. 1, pp. 1–15, 2021.
- “Lung segmentation in chest radiographs using anatomical atlases with nonrigid registration,” IEEE Trans. Medical Imaging, vol. 33, no. 2, pp. 577–590, 2014.
- “A benchmark for comparison of dental radiography analysis algorithms,” Medical Image Anal., vol. 31, pp. 63–76, 2016.
- “Integrating spatial configuration into heatmap regression based cnns for landmark localization,” Medical Image Anal., vol. 54, pp. 207–219, 2019.
- “Unet++: A nested u-net architecture for medical image segmentation,” CoRR, vol. abs/1807.10165, 2018.
- “U-net: Convolutional networks for biomedical image segmentation,” CoRR, vol. abs/1505.04597, 2015.
- “Rethinking atrous convolution for semantic image segmentation,” CoRR, vol. abs/1706.05587, 2017.
- “Fully automatic system for accurate localisation and analysis of cephalometric landmarks in lateral cephalograms,” Scientific Reports, vol. 6, no. 1, pp. 33581, 2016.
- “Integrating geometric configuration and appearance information into a unified framework for anatomical landmark localization,” Medical Image Analysis, vol. 43, 09 2017.
- Roberto Di Via (3 papers)
- Matteo Santacesaria (35 papers)
- Francesca Odone (21 papers)
- Vito Paolo Pastore (10 papers)