MV-Swin-T: Mammogram Classification with Multi-view Swin Transformer (2402.16298v1)
Abstract: Traditional deep learning approaches for breast cancer classification has predominantly concentrated on single-view analysis. In clinical practice, however, radiologists concurrently examine all views within a mammography exam, leveraging the inherent correlations in these views to effectively detect tumors. Acknowledging the significance of multi-view analysis, some studies have introduced methods that independently process mammogram views, either through distinct convolutional branches or simple fusion strategies, inadvertently leading to a loss of crucial inter-view correlations. In this paper, we propose an innovative multi-view network exclusively based on transformers to address challenges in mammographic image classification. Our approach introduces a novel shifted window-based dynamic attention block, facilitating the effective integration of multi-view information and promoting the coherent transfer of this information between views at the spatial feature map level. Furthermore, we conduct a comprehensive comparative analysis of the performance and effectiveness of transformer-based models under diverse settings, employing the CBIS-DDSM and Vin-Dr Mammo datasets. Our code is publicly available at https://github.com/prithuls/MV-Swin-T
- American Cancer Society, “Breast cancer facts & figures 2019–2020,” Am Cancer Soc, pp. 1–44, 2019.
- “Hypercomplex neural architectures for multi-view breast cancer classification,” arXiv preprint arXiv:2204.05798, 2022.
- “Multi-view convolutional neural networks for mammographic image classification,” IEEE Access, vol. 7, pp. 126273–126282, 2019.
- “Triage of 2d mammographic images using multi-view multi-task convolutional neural networks,” ACM Transactions on Computing for Healthcare, vol. 2, no. 3, pp. 1–24, 2021.
- “Deep learning models for classifying mammogram exams containing unregistered multi-view images and segmentation maps of lesions,” Deep learning for medical image analysis, pp. 321–339, 2017.
- V Sridevi and J Abdul Samath, “A combined deep cnn-lasso regression feature fusion and classification of mlo and cc view mammogram image,” International Journal of System Assurance Engineering and Management, pp. 1–11, 2023.
- “Multi-view feature fusion based four views model for mammogram classification using convolutional neural network,” IEEE Access, vol. 7, pp. 165724–165733, 2019.
- “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
- “Multi-view analysis of unregistered medical images using cross-view transformers,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part III 24. Springer, 2021, pp. 104–113.
- “Transformers improve breast cancer diagnosis from unregistered multi-view mammograms,” Diagnostics, vol. 12, no. 7, pp. 1549, 2022.
- “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 10012–10022.
- “Momminet-v2: Mammographic multi-view mass identification networks,” Medical Image Analysis, vol. 73, pp. 102204, 2021.
- “A curated mammography data set for use in computer-aided detection and diagnosis research,” Scientific data, vol. 4, no. 1, pp. 1–9, 2017.
- “Vindr-mammo: A large-scale benchmark dataset for computer-aided diagnosis in full-field digital mammography,” Scientific Data, vol. 10, no. 1, pp. 277, 2023.
- “Connectedunets++: Mass segmentation from whole mammographic images,” in International Symposium on Visual Computing. Springer, 2022, pp. 419–430.
- “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.