Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Conditional Graph Information Bottleneck for Molecular Relational Learning (2305.01520v2)

Published 29 Apr 2023 in q-bio.MN and cs.LG

Abstract: Molecular relational learning, whose goal is to learn the interaction behavior between molecular pairs, got a surge of interest in molecular sciences due to its wide range of applications. Recently, graph neural networks have recently shown great success in molecular relational learning by modeling a molecule as a graph structure, and considering atom-level interactions between two molecules. Despite their success, existing molecular relational learning methods tend to overlook the nature of chemistry, i.e., a chemical compound is composed of multiple substructures such as functional groups that cause distinctive chemical reactions. In this work, we propose a novel relational learning framework, called CGIB, that predicts the interaction behavior between a pair of graphs by detecting core subgraphs therein. The main idea is, given a pair of graphs, to find a subgraph from a graph that contains the minimal sufficient information regarding the task at hand conditioned on the paired graph based on the principle of conditional graph information bottleneck. We argue that our proposed method mimics the nature of chemical reactions, i.e., the core substructure of a molecule varies depending on which other molecule it interacts with. Extensive experiments on various tasks with real-world datasets demonstrate the superiority of CGIB over state-of-the-art baselines. Our code is available at https://github.com/Namkyeong/CGIB.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (58)
  1. Deep variational information bottleneck. arXiv preprint arXiv:1612.00410, 2016.
  2. Subgraph neural networks. Advances in Neural Information Processing Systems, 33:8017–8029, 2020.
  3. Simgnn: A neural network approach to fast graph similarity computation. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, pp.  384–392, 2019.
  4. Learning-based efficient graph similarity computation via multi-scale convolutional set matching. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp.  3219–3226, 2020.
  5. Book, G. Compendium of chemical terminology. International Union of Pure and Applied Chemistry, 528, 2014.
  6. Bunke, H. On a relation between graph edit distance and maximum common subgraph. Pattern recognition letters, 18(8):689–694, 1997.
  7. A graph distance metric based on the maximal common subgraph. Pattern recognition letters, 19(3-4):255–259, 1998.
  8. Drug-drug adverse effect prediction with graph co-attention. arXiv preprint arXiv:1905.00534, 2019.
  9. Delaney, J. S. Esol: estimating aqueous solubility directly from molecular structure. Journal of chemical information and computer sciences, 44(3):1000–1005, 2004.
  10. Hypergraph neural networks. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp.  3558–3565, 2019.
  11. Neural message passing for quantum chemistry. In International conference on machine learning, pp. 1263–1272. PMLR, 2017.
  12. Tradeoffs in data augmentation: An empirical study. In International Conference on Learning Representations, 2020.
  13. Mathematical correlations for describing solute transfer into functionalized alkane solvents containing hydroxyl, ether, ester or ketone solvents. Fluid phase equilibria, 298(1):48–53, 2010.
  14. Learning deep representations by mutual information estimation and maximization. arXiv preprint arXiv:1808.06670, 2018.
  15. Hynes, J. T. Chemical reaction dynamics in solution. Annual Review of Physical Chemistry, 36(1):573–597, 1985.
  16. Categorical reparameterization with gumbel-softmax. arXiv preprint arXiv:1611.01144, 2016.
  17. Jerry, M. Advanced organic chemistry: reactions, mechanisms and structure, 1992.
  18. Experimental database of optical properties of organic compounds. Scientific data, 7(1):1–6, 2020.
  19. Deep learning optical spectroscopy based on experimental database: Potential applications to molecular design. JACS Au, 1(4):427–438, 2021.
  20. Predicting potential drug-drug interactions on topological and semantic similarity features using statistical learning. PloS one, 13(5):e0196865, 2018.
  21. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
  22. Graph matching networks for learning the similarity of graph structured objects. In International conference on machine learning, pp. 3835–3845. PMLR, 2019.
  23. Delfos: deep learning model for prediction of solvation free energies in generic organic solvents. Chemical science, 10(36):8306–8315, 2019.
  24. Hierarchical graph matching networks for deep graph similarity learning. 2019.
  25. The concrete distribution: A continuous relaxation of discrete random variables. arXiv preprint arXiv:1611.00712, 2016.
  26. Minnesota solvation database (mnsol) version 2012. 2020.
  27. Subgraph pattern neural networks for high-order graph evolution prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018.
  28. Interpretable and generalizable graph learning via stochastic attention mechanism. In International Conference on Machine Learning, pp. 15524–15543. PMLR, 2022.
  29. Freesolv: a database of experimental and calculated hydration free energies, with input files. Journal of computer-aided molecular design, 28(7):711–720, 2014.
  30. Estimation of solvation quantities from experimental thermodynamic data: Development of the comprehensive compsol databank for pure and mixed solutes. Journal of Physical and Chemical Reference Data, 46(3):033102, 2017.
  31. Ssi–ddi: substructure–substructure interactions for drug–drug interaction prediction. Briefings in Bioinformatics, 22(6):bbab133, 2021.
  32. Disentangled information bottleneck. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp.  9285–9293, 2021.
  33. Chemically interpretable graph interaction network for prediction of pharmacokinetic properties of drug-like molecules. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp.  873–880, 2020.
  34. Deepsynergy: predicting anti-cancer drug synergy with deep learning. Bioinformatics, 34(9):1538–1546, 2018.
  35. Fluorine in medicinal chemistry. Chemical Society Reviews, 37(2):320–330, 2008.
  36. Reichardt, C. Empirical parameters of the polarity of solvents. Angewandte Chemie International Edition in English, 4(1):29–40, 1965.
  37. A unified view of relational deep learning for drug pair scoring. arXiv preprint arXiv:2111.02916, 2021.
  38. Effect of solvent hydrogen bonding on the photophysical properties of intramolecular charge transfer probe trans-ethyl p-(dimethylamino) cinamate and its derivative. Spectrochimica Acta Part A: Molecular and Biomolecular Spectroscopy, 73(4):630–636, 2009.
  39. Graph structure learning with variational information bottleneck. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp.  4165–4174, 2022.
  40. Contrastive multiview coding. In European conference on computer vision, pp.  776–794. Springer, 2020.
  41. The information bottleneck method. arXiv preprint physics/0004057, 2000.
  42. Graph attention networks. 2017.
  43. Deep graph infomax. arXiv preprint arXiv:1809.10341, 2018.
  44. Transfer learning for solvation free energies: From quantum chemistry to experiments. Chemical Engineering Journal, 418:129307, 2021.
  45. Drug—drug interaction through molecular structure similarity analysis. Journal of the American Medical Informatics Association, 19(6):1066–1074, 2012.
  46. Order matters: Sequence to sequence for sets. arXiv preprint arXiv:1511.06391, 2015.
  47. Multi-view graph contrastive representation learning for drug-drug interaction prediction. In Proceedings of the Web Conference 2021, pp.  2921–2933, 2021.
  48. Graph information bottleneck. Advances in Neural Information Processing Systems, 33:20437–20448, 2020.
  49. How powerful are graph neural networks? arXiv preprint arXiv:1810.00826, 2018.
  50. Neural network-based graph embedding for cross-platform binary code similarity detection. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, pp.  363–376, 2017.
  51. Graph contrastive learning with augmentations. Advances in Neural Information Processing Systems, 33:5812–5823, 2020.
  52. Graph information bottleneck for subgraph recognition. arXiv preprint arXiv:2010.05563, 2020.
  53. Improving subgraph recognition with variational graph information bottleneck. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  19396–19405, 2022.
  54. Predicting potential drug-drug interactions by integrating chemical, biological, phenotypic and network data. BMC bioinformatics, 18(1):1–12, 2017.
  55. H2mn: Graph similarity learning with hierarchical hypergraph matching networks. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp.  2274–2284, 2021.
  56. Learning with hypergraphs: Clustering, classification, and embedding. Advances in neural information processing systems, 19, 2006.
  57. Graph contrastive learning with adaptive augmentation. In Proceedings of the Web Conference 2021, pp.  2069–2080, 2021.
  58. Biosnap datasets: Stanford biomedical network dataset collection. Note: http://snap. stanford. edu/biodata Cited by, 5(1), 2018.
Citations (16)

Summary

  • The paper presents the CGIB framework that conditionally extracts core molecular subgraphs to enhance predictive accuracy for interaction tasks.
  • It employs the Information Bottleneck principle to focus on task-relevant substructures while pruning non-essential graph components.
  • Experimental results demonstrate CGIB’s robustness in both transductive and inductive settings, outperforming strong baseline models.

Conditional Graph Information Bottleneck for Molecular Relational Learning

The paper "Conditional Graph Information Bottleneck for Molecular Relational Learning" introduces the Conditional Graph Information Bottleneck (CGIB) framework, a novel approach to molecular relational learning that aims to accurately predict interaction behaviors between molecular pairs by identifying crucial substructures within them. This method enhances the predictive ability of molecular interaction models by focusing on segments of molecular graphs that are most relevant for a given task, leveraging concepts from the Information Bottleneck (IB) theory.

Overview of Molecular Relational Learning and Current Challenges

Molecular relational learning models the interactions within molecular pairs, translating these interactions into graph representations where atoms are nodes, and bonds are edges. Graph Neural Networks (GNNs) have previously been used in this field, but they often fail to account adequately for substructures like functional groups which define the chemical reactions of molecules. For instance, in chemical contexts such as drug interactions or solubility predictions, only certain substructures may influence the reaction and need to be detected dynamically based on interacting molecules. The new CGIB paradigm directly addresses this by allowing the core substructure of a molecule to adapt based on its interaction partner.

Conditional Graph Information Bottleneck Framework

CGIB builds upon the IB principle to conditionally learn core subgraphs that remain invariant to the irrelevant molecular components. The methodology involves maximizing the mutual information between the core subgraph and the prediction target while minimizing the shared information between the original graph and its contracted subgraph, conditioned on its interaction pair. This approach dynamically discerns the task-specific core substructure by exploiting the chain rule of mutual information, integrating Gaussian noise to prune less critical parts of the graph representation.

Experimental Validation and Results

Empirical validation on various real-world datasets demonstrates the superiority of CGIB over existing models, particularly against strong baselines such as CIGIN for molecular interaction tasks, and SSI-DDI for drug-drug interactions. The datasets involved molecular properties like solubility and drug interactions, where CGIB consistently enhances predictive performance. Notably, the research highlights the robustness of CGIB in both transductive and inductive settings, showcasing its generalization capabilities in scenarios where interaction pairs are entirely new graph instances not encountered during training.

Practical and Theoretical Implications

CGIB's ability to adaptively determine the core substructure of interacting molecules aligns closely with practical needs in molecular science, such as drug design, by offering a mechanistic insight into chemical interactions. Theoretically, CGIB extends the applicability of the Information Bottleneck principle to relational learning tasks, opening avenues for future research to explore conditional adaptations within diverse domains where relationships between entities must be contextually evaluated.

Future Directions

Looking ahead, the core insights from CGIB could be extended to more complex chemical systems or to other domains requiring relational mappings, such as social networks or biological systems. Additionally, examining different types of GNN architectures within the CGIB framework or integrating domain-specific knowledge regarding chemical properties could further enhance this method's utility.

By prioritizing task-specific molecular substructures in a relational context, CGIB provides a promising direction for more explainable and efficient molecular relational learning. This forward step can significantly impact specialized fields requiring precise interaction predictions, such as pharmaceuticals and materials science.

X Twitter Logo Streamline Icon: https://streamlinehq.com