Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pathomic Fusion: An Integrated Framework for Fusing Histopathology and Genomic Features for Cancer Diagnosis and Prognosis (1912.08937v3)

Published 18 Dec 2019 in cs.CV, q-bio.GN, and q-bio.TO

Abstract: Cancer diagnosis, prognosis, and therapeutic response predictions are based on morphological information from histology slides and molecular profiles from genomic data. However, most deep learning-based objective outcome prediction and grading paradigms are based on histology or genomics alone and do not make use of the complementary information in an intuitive manner. In this work, we propose Pathomic Fusion, an interpretable strategy for end-to-end multimodal fusion of histology image and genomic (mutations, CNV, RNA-Seq) features for survival outcome prediction. Our approach models pairwise feature interactions across modalities by taking the Kronecker product of unimodal feature representations and controls the expressiveness of each representation via a gating-based attention mechanism. Following supervised learning, we are able to interpret and saliently localize features across each modality, and understand how feature importance shifts when conditioning on multimodal input. We validate our approach using glioma and clear cell renal cell carcinoma datasets from the Cancer Genome Atlas (TCGA), which contains paired whole-slide image, genotype, and transcriptome data with ground truth survival and histologic grade labels. In a 15-fold cross-validation, our results demonstrate that the proposed multimodal fusion paradigm improves prognostic determinations from ground truth grading and molecular subtyping, as well as unimodal deep networks trained on histology and genomic data alone. The proposed method establishes insight and theory on how to train deep networks on multimodal biomedical data in an intuitive manner, which will be useful for other problems in medicine that seek to combine heterogeneous data streams for understanding diseases and predicting response and resistance to treatment.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Richard J. Chen (28 papers)
  2. Ming Y. Lu (23 papers)
  3. Jingwen Wang (34 papers)
  4. Drew F. K. Williamson (24 papers)
  5. Scott J. Rodig (3 papers)
  6. Neal I. Lindeman (1 paper)
  7. Faisal Mahmood (53 papers)
Citations (341)

Summary

An Analysis of Pathomic Fusion for Cancer Diagnosis and Prognosis

The paper "Pathomic Fusion: An Integrated Framework for Fusing Histopathology and Genomic Features for Cancer Diagnosis and Prognosis" offers a novel approach to integrate histopathology and genomic data for improved cancer diagnosis, prognosis, and survival prediction. The authors propose Pathomic Fusion, a comprehensive framework that utilizes deep learning techniques to leverage the complementary information from histological images and genomic data in an end-to-end multimodal learning paradigm. This strategy aims to address the limitations of current predictive models that typically rely on a single modality, either genomics or histology.

The primary methodological advancement presented in this paper is the use of the Kronecker product for modeling interactions across different modalities. This approach, combined with a gating-based attention mechanism, captures the intricate relationships between histology and genomic features more effectively than traditional concatenation methods. Additionally, the paper employs a triad of neural network architectures: Convolutional Neural Networks (CNNs) for extracting features from histopathology images, Graph Convolutional Networks (GCNs) for learning cell graph features from histological tissues, and Self-Normalizing Networks (SNNs) for processing genomic data. This robust combination enhances the model's ability to perform survival analysis and classification tasks.

The authors validate their framework using datasets from The Cancer Genome Atlas (TCGA) on glioma and clear cell renal cell carcinoma (CCRCC), demonstrating significant improvements over existing methods in stratifying patients by risk and predicting survival outcomes. Notably, Pathomic Fusion's application resulted in a concordance index (c-Index) of 0.826 for glioma, surpassing prior multimodal benchmarks. The model also provided more detailed and statistically significant patient stratification compared to traditional prognostic methods such as the WHO grading system and the Fuhrman Grading System.

An essential aspect of the paper is the multimodal interpretability of the Pathomic Fusion framework. This capability allows for a detailed understanding of how different features across modalities contribute to the predictions, highlighting prognostic markers relevant to glioma and CCRCC. The integration of interpretability techniques such as Grad-CAM and Integrated Gradients facilitates a transparent analysis of the features considered critical by the model.

The practical implications of this research are substantial. The ability of Pathomic Fusion to integrate heterogeneous data types promises enhanced diagnostic accuracy and patient stratification, potentially leading to more personalized treatment approaches in oncology. Theoretically, the advancements in multimodal fusions, such as the application of Kronecker products and attention mechanisms, can be extended to other domains where complex interactions between multiple data types exist.

Future developments could focus on refining the model's ability to handle missing or imbalanced data across modalities, which remains a challenge in clinical datasets. Moreover, extending this framework to other cancer types or integrating additional data modalities, such as clinical records or proteomics, could further improve predictive performance and applicability in personalized medicine.

By bridging the gap between histopathology and genomic features, Pathomic Fusion exemplifies the potential of multimodal deep learning in complex medical tasks, setting a precedent for future research in integrative cancer diagnostics and prognostics.