Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
GPT-4o
Gemini 2.5 Pro Pro
o3 Pro
GPT-4.1 Pro
DeepSeek R1 via Azure Pro
2000 character limit reached

Adapting Differential Molecular Representation with Hierarchical Prompts for Multi-label Property Prediction (2405.18724v2)

Published 29 May 2024 in q-bio.QM, cs.AI, and cs.LG

Abstract: Accurate prediction of molecular properties is crucial in drug discovery. Traditional methods often overlook that real-world molecules typically exhibit multiple property labels with complex correlations. To this end, we propose a novel framework, HiPM, which stands for hierarchical prompted molecular representation learning framework. HiPM leverages task-aware prompts to enhance the differential expression of tasks in molecular representations and mitigate negative transfer caused by conflicts in individual task information. Our framework comprises two core components: the Molecular Representation Encoder (MRE) and the Task-Aware Prompter (TAP). MRE employs a hierarchical message-passing network architecture to capture molecular features at both the atom and motif levels. Meanwhile, TAP utilizes agglomerative hierarchical clustering algorithm to construct a prompt tree that reflects task affinity and distinctiveness, enabling the model to consider multi-granular correlation information among tasks, thereby effectively handling the complexity of multi-label property prediction. Extensive experiments demonstrate that HiPM achieves state-of-the-art performance across various multi-label datasets, offering a novel perspective on multi-label molecular representation learning.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Molecular property prediction: recent trends in the era of artificial intelligence. Drug Discovery Today: Technologies, 32-33:29–36, 2019.
  2. Molclr: Molecular contrastive learning of representations via graph neural networks. Nature Machine Intelligence, 4:279–287, 2022.
  3. Molfescue: enhancing molecular property prediction in data-limited and imbalanced contexts using few-shot and contrastive learning. Bioinformatics, 40:btae118, 2024.
  4. Aegnn-m:a 3d graph-spatial co-representation model for molecular property prediction. IEEE Journal of Biomedical and Health Informatics, pages 1–9, 2024.
  5. Property-guided few-shot learning for molecular property prediction with dual-view encoder and relation graph learning network. IEEE Journal of Biomedical and Health Informatics, pages 1–12, 2024.
  6. Cashman J. The mechanisms of action of nsaids in analgesia. Drugs, 52:13–23, 1996.
  7. Chemberta: Large-scale self-supervised pretraining for molecular property prediction, 2020.
  8. 3d infomax improves gnns for molecular property prediction. In International Conference on Machine Learning, 2022.
  9. Geometry-enhanced molecular representation learning for property prediction. Nature Machine Intelligence, 4:127–134, 2022.
  10. Property-aware relation networks for few-shot molecular property prediction. Advances in Neural Information Processing Systems, 34:17441–17454, 2021.
  11. Transfer learning with graph neural networks for improved molecular property prediction in the multi-fidelity setting. Nature Communications, 15:1517, 2024.
  12. Fast and effective molecular property prediction with transferability map. Communications Chemistry, 7:85, 2024.
  13. A review on multi-label learning algorithms. IEEE Transactions on Knowledge and Data Engineering, 26:1819–1837, 2014.
  14. The emerging trends of multi-label learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44:7955–7974, 2022.
  15. Gradient surgery for multi-task learning. In Proceedings of the 34th International Conference on Neural Information Processing Systems, 2020.
  16. Hierarchical prompt learning for multi-task learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
  17. Language models are few-shot learners. In Advances in Neural Information Processing Systems, 2020.
  18. Zero-shot text-to-image generation. In Proceedings of the 38th International Conference on Machine Learning, 2021.
  19. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput. Surv., 55, 2023.
  20. Gppt: Graph pre-training and prompt tuning to generalize graph neural networks. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2022.
  21. All in one: Multi-task prompting for graph neural networks. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2023.
  22. Graphprompt: Unifying pre-training and downstream tasks for graph neural networks. In Proceedings of the ACM Web Conference, 2023.
  23. Universal prompt tuning for graph neural networks, 2024.
  24. Moleculenet: A benchmark for molecular machine learning, 2018.
  25. Neural message passing for quantum chemistry. In Proceedings of the 34th International Conference on Machine Learning, 2017.
  26. Communicative representation learning on attributed molecular graphs. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, 2020.
  27. Learning multimodal graph-to-graph translation for molecular optimization, 2019.
  28. Advanced graph and sequence neural networks for molecular property prediction and drug discovery. Bioinformatics, 38:2579–2586, 2022.
  29. Exploiting cloze-questions for few-shot text classification and natural language inference. In Conference of the European Chapter of the Association for Computational Linguistics, 2020.
  30. Gpt understands, too. AI Open, 2023.
  31. Moltailor: Tailoring chemical molecular representation to specific tasks via text prompts. In AAAI Conference on Artificial Intelligence, 2024.
  32. Knowledge graph-enhanced molecular contrastive learning with functional prompt. Nature Machine Intelligence, 5:542–553, 2023.
  33. David Weininger. Smiles, a chemical language and information system. 1. introduction to methodology and encoding rules. Journal of Chemical Information and Computer Sciences, 28:31–36, 1988.
  34. Molecular joint representation learning via multi-modal information of smiles and graphs. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 20:3044–3055, 2023.
  35. Geometric deep learning on molecular representations. Nature Machine Intelligence, 3:1023–1032, 2021.
  36. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017.
  37. Conflict-averse gradient descent for multi-task learning. In Advances in Neural Information Processing Systems, 2021.
  38. Analyzing learned molecular representations for property prediction. Journal of Chemical Information and Modeling, 59:3370–3388, 2019.
  39. FP-GNN: a versatile deep learning architecture for enhanced molecular property prediction. Briefings in Bioinformatics, 23:bbac408, 2022.
  40. Cross-dependent graph neural networks for molecular property prediction. Bioinformatics, 38:2003–2009, 2022.
  41. Self-supervised graph transformer on large-scale molecular data. In Proceedings of the 34th International Conference on Neural Information Processing Systems, 2020.
  42. Motif-based graph self-supervised learning for molecular property prediction. In Neural Information Processing Systems, 2021.
  43. Hierarchical molecular graph self-supervised learning for property prediction. Communications Chemistry, 6, 2023.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com