Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 22 tok/s Pro
GPT-4o 93 tok/s Pro
Kimi K2 205 tok/s Pro
GPT OSS 120B 426 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

TNF: Tri-branch Neural Fusion for Multimodal Medical Data Classification (2403.01802v3)

Published 4 Mar 2024 in cs.CV

Abstract: This paper presents a Tri-branch Neural Fusion (TNF) approach designed for classifying multimodal medical images and tabular data. It also introduces two solutions to address the challenge of label inconsistency in multimodal classification. Traditional methods in multi-modality medical data classification often rely on single-label approaches, typically merging features from two distinct input modalities. This becomes problematic when features are mutually exclusive or labels differ across modalities, leading to reduced accuracy. To overcome this, our TNF approach implements a tri-branch framework that manages three separate outputs: one for image modality, another for tabular modality, and a third hybrid output that fuses both image and tabular data. The final decision is made through an ensemble method that integrates likelihoods from all three branches. We validate the effectiveness of TNF through extensive experiments, which illustrate its superiority over traditional fusion and ensemble methods in various convolutional neural networks and transformer-based architectures across multiple datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (45)
  1. Multimodal biomedical AI. Nature Medicine, 28(9):1773–1784, 2022.
  2. Use of multi-modal data and machine learning to improve cardiovascular disease care. Frontiers in Cardiovascular Medicine, 9:840262, 2022.
  3. Alzheimer’s Association. Medical tests for diagnosing Alzheimer’s. https://www.alz.org/alzheimers-dementia/diagnosis/medical_tests, 2023. Accessed: 2023-11-09.
  4. Multimodal machine learning: A survey and taxonomy. IEEE transactions on pattern analysis and machine intelligence, 41(2):423–443, 2018.
  5. The national Alzheimer’s coordinating center (NACC) database: an Alzheimer disease database. Alzheimer Disease & Associated Disorders, 18(4):270–277, 2004.
  6. The power of ensembles for active learning in image classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 9368–9377, 2018.
  7. A short note about Kinetics-600. arXiv preprint arXiv:1808.01340, 2018.
  8. The Matthews correlation coefficient (MCC) is more reliable than balanced accuracy, bookmaker informedness, and markedness in two-class confusion matrix evaluation. BioData mining, 14(1):1–22, 2021.
  9. Weighted feature fusion of convolutional neural network and graph attention network for hyperspectral image classification. IEEE Transactions on Image Processing, 31:1559–1572, 2022.
  10. Bruce Fischl. Freesurfer. Neuroimage, 62(2):774–781, 2012.
  11. Video-based surgical skill assessment using 3D convolutional neural networks. International journal of computer assisted radiology and surgery, 14:1217–1225, 2019.
  12. Revisiting deep learning models for tabular data. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 18932–18943. Curran Associates, Inc., 2021.
  13. Best of both worlds: Multimodal contrastive learning with tabular and imaging data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 23924–23935, June 2023.
  14. End-to-end learning of fused image and non-image features for improved breast cancer classification from MRI. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3294–3303, 2021.
  15. Improved multimodal fusion for small datasets with auxiliary supervision. arXiv preprint arXiv:2304.00379, 2023.
  16. PENet—a scalable deep-learning model for automated diagnosis of pulmonary embolism using volumetric CT imaging. NPJ digital medicine, 3(1):61, 2020.
  17. Deep residual learning for image recognition. In IEEE Conference on Computer Vision & Pattern Recognition, pages 770–778, 2016.
  18. MMTM: Multimodal transfer module for CNN fusion. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 13289–13299, 2020.
  19. An ensemble of fine-tuned convolutional neural networks for medical image classification. IEEE journal of biomedical and health informatics, 21(1):31–40, 2016.
  20. Multimodal fusion with co-attention mechanism. In 2020 IEEE 23rd International Conference on Information Fusion (FUSION), pages 1–8. IEEE, 2020.
  21. Decoupled weight decay regularization. In International Conference on Learning Representations, 2018.
  22. A unified approach to interpreting model predictions. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 4765–4774. Curran Associates, Inc., 2017.
  23. Predicting cancer outcomes from histology and genomics using convolutional networks. Proceedings of the National Academy of Sciences, 115(13):E2970–E2979, 2018.
  24. Clinical text data categorization and feature extraction using medical-fissure algorithm and Neg-Seq algorithm. Computational Intelligence and Neuroscience, 2022(5759521), 2022.
  25. AdaMML: Adaptive multi-modal learning for efficient video recognition. In Proceedings of the IEEE/CVF international conference on computer vision, pages 7576–7585, 2021.
  26. A review of the application of multi-modal deep learning in medicine: bibliometrics and future directions. International Journal of Computational Intelligence Systems, 16(1):44, 2023.
  27. Multimodal deep learning for Alzheimer’s disease dementia assessment. Nature communications, 13(1):3404, 2022.
  28. Lior Rokach. Pattern classification using ensemble methods, volume 75. World Scientific, 2010.
  29. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision, pages 618–626, 2017.
  30. A survey of multimodal information fusion for smart healthcare: Mapping the journey from data to wisdom. Information Fusion, page 102040, 2023.
  31. Learning important features through propagating activation differences. In International conference on machine learning, pages 3145–3153. PMLR, 2017.
  32. 3D deep learning on medical images: a review. Sensors, 20(18):5097, 2020.
  33. Deep learning in mental health outcome research: a scoping review. Translational Psychiatry, 10(1):116, 2020.
  34. Multimodal self-supervised learning for medical image analysis. In International conference on information processing in medical imaging, pages 661–673. Springer, 2021.
  35. Multimodal research in vision and language: A review of current and emerging trends. Information Fusion, 77, 2021.
  36. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  37. Medical data classification assisted by machine learning strategy. Journal of Artificial Intelligence Research, 2022(9699612), 2022.
  38. Multimodal token fusion for vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12186–12195, 2022.
  39. Multimodal end-to-end autonomous driving. IEEE Transactions on Intelligent Transportation Systems, 23(1):537–547, 2020.
  40. Deep correlational learning for survival prediction from multi-modality data. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 406–414. Springer, 2017.
  41. Multimodal skin lesion classification using deep learning. Experimental dermatology, 27(11):1261–1267, 2018.
  42. Deep learning of brain lesion patterns and user-defined clinical and MRI features for predicting conversion to multiple sclerosis from clinically isolated syndrome. Computer Methods in Biomechanics and Biomedical Engineering: Imaging & Visualization, 7(3):250–259, 2019.
  43. Information fusion for classification of hyperspectral and LiDAR data using IP-CNN. IEEE Transactions on Geoscience and Remote Sensing, 60:1–12, 2021.
  44. Radfusion: Benchmarking performance and fairness for multimodal pulmonary embolism detection from CT and EHR. arXiv preprint arXiv:2111.11665, 2021.
  45. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society Series B: Statistical Methodology, 67(2):301–320, 2005.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 0 likes.

Upgrade to Pro to view all of the tweets about this paper: