Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pretraining Strategy for Neural Potentials (2402.15921v2)

Published 24 Feb 2024 in cs.LG and physics.chem-ph

Abstract: We propose a mask pretraining method for Graph Neural Networks (GNNs) to improve their performance on fitting potential energy surfaces, particularly in water systems. GNNs are pretrained by recovering spatial information related to masked-out atoms from molecules, then transferred and finetuned on atomic forcefields. Through such pretraining, GNNs learn meaningful prior about structural and underlying physical information of molecule systems that are useful for downstream tasks. From comprehensive experiments and ablation studies, we show that the proposed method improves the accuracy and convergence speed compared to GNNs trained from scratch or using other pretraining techniques such as denoising. On the other hand, our pretraining method is suitable for both energy-centric and force-centric GNNs. This approach showcases its potential to enhance the performance and data efficiency of GNNs in fitting molecular force fields.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Unke, O. T.; Chmiela, S.; Sauceda, H. E.; Gastegger, M.; Poltavsky, I.; Schütt, K. T.; Tkatchenko, A.; Müller, K.-R. Machine Learning Force Fields. 2021
  2. Chmiela, S.; Tkatchenko, A.; Sauceda, H. E.; Poltavsky, I.; Schütt, K. T.; Müller, K.-R. Machine learning of accurate energy-conserving molecular force fields. Science Advances 2017, 3
  3. Bartók, A. P.; De, S.; Poelking, C.; Bernstein, N.; Kermode, J. R.; Csányi, G.; Ceriotti, M. Machine learning unifies the modeling of materials and molecules. Science Advances 2017, 3
  4. Duvenaud, D. K.; Maclaurin, D.; Iparraguirre, J.; Bombarell, R.; Hirzel, T.; Aspuru-Guzik, A.; Adams, R. P. Convolutional Networks on Graphs for Learning Molecular Fingerprints. Advances in Neural Information Processing Systems. 2015
  5. Schütt, K. T.; Kindermans, P.-J.; Sauceda, H. E.; Chmiela, S.; Tkatchenko, A.; Müller, K.-R. SchNet: A continuous-filter convolutional neural network for modeling quantum interactions. 2017
  6. Gilmer, J.; Schoenholz, S. S.; Riley, P. F.; Vinyals, O.; Dahl, G. E. Neural Message Passing for Quantum Chemistry. Proceedings of the 34th International Conference on Machine Learning. 2017; pp 1263–1272
  7. Klicpera, J.; Groß, J.; Günnemann, S. Directional Message Passing for Molecular Graphs. International Conference on Learning Representations. 2020
  8. Thomas, N.; Smidt, T.; Kearnes, S.; Yang, L.; Li, L.; Kohlhoff, K.; Riley, P. Tensor field networks: Rotation-and translation-equivariant neural networks for 3d point clouds. arXiv preprint arXiv:1802.08219 2018,
  9. Anderson, B.; Hy, T. S.; Kondor, R. Cormorant: Covariant molecular neural networks. Advances in neural information processing systems 2019, 32
  10. Jing, B.; Eismann, S.; Suriana, P.; Townshend, R. J. L.; Dror, R. Learning from Protein Structure with Geometric Vector Perceptrons. International Conference on Learning Representations. 2021
  11. Villar, S.; Hogg, D. W.; Storey-Fisher, K.; Yao, W.; Blum-Smith, B. Scalars are universal: Equivariant machine learning, structured like classical physics. Advances in Neural Information Processing Systems. 2021
  12. Hu*, W.; Liu*, B.; Gomes, J.; Zitnik, M.; Liang, P.; Pande, V.; Leskovec, J. Strategies for Pre-training Graph Neural Networks. International Conference on Learning Representations. 2020
  13. Wang, Y.; Wang, J.; Cao, Z.; Farimani, A. B. MolCLR: Molecular Contrastive Learning of Representations via Graph Neural Networks. 2021
  14. Liu, Y.; Jin, M.; Pan, S.; Zhou, C.; Zheng, Y.; Xia, F.; Yu, P. Graph self-supervised learning: A survey. IEEE Transactions on Knowledge and Data Engineering 2022,
  15. Krishnan, R.; Rajpurkar, P.; Topol, E. J. Self-supervised learning in medicine and healthcare. Nature Biomedical Engineering 2022, 1–7
  16. Cao, Z.; Magar, R.; Wang, Y.; Farimani, A. B. MOFormer: Self-Supervised Transformer model for Metal-Organic Framework Property Prediction. arXiv preprint arXiv:2210.14188 2022,
  17. Zhang, S.; Hu, Z.; Subramonian, A.; Sun, Y. Motif-driven contrastive learning of graph representations. arXiv preprint arXiv:2012.12533 2020,
  18. Wang, Y.; Magar, R.; Liang, C.; Barati Farimani, A. Improving Molecular Contrastive Learning via Faulty Negative Mitigation and Decomposed Fragment Contrast. Journal of Chemical Information and Modeling 2022,
  19. Stärk, H.; Beaini, D.; Corso, G.; Tossou, P.; Dallago, C.; Günnemann, S.; Liò, P. 3d infomax improves gnns for molecular property prediction. International Conference on Machine Learning. 2022; pp 20479–20502
  20. Liu, S.; Wang, H.; Liu, W.; Lasenby, J.; Guo, H.; Tang, J. Pre-training Molecular Graph Representation with 3D Geometry. International Conference on Learning Representations. 2022
  21. Zaidi, S.; Schaarschmidt, M.; Martens, J.; Kim, H.; Teh, Y. W.; Sanchez-Gonzalez, A.; Battaglia, P.; Pascanu, R.; Godwin, J. Pre-training via Denoising for Molecular Property Prediction. arXiv preprint arXiv:2206.00133 2022,
  22. Liu, S.; Guo, H.; Tang, J. Molecular geometry pretraining with se (3)-invariant denoising distance matching. arXiv preprint arXiv:2206.13602 2022,
  23. Zhou, G.; Gao, Z.; Ding, Q.; Zheng, H.; Xu, H.; Wei, Z.; Zhang, L.; Ke, G. Uni-Mol: A Universal 3D Molecular Representation Learning Framework. The Eleventh International Conference on Learning Representations. 2023
  24. Jiang, D.; Zhenxing Wu, C.-Y. H.; Chen, G.; Liao, B.; Wang, Z.; Shen, C.; Cao, D.; Wu, J.; Hou, T. Could graph neural networks learn better molecular representation for drug discovery? A comparison study of descriptor-based and graph-based models. Journal of Cheminformatics 2021,
  25. Rong, Y.; Bian, Y.; Xu, T.; Xie, W.; Wei, Y.; Huang, W.; Huang, J. Self-Supervised Graph Transformer on Large-Scale Molecular Data. 2020,
  26. Satorras, V. G.; Hoogeboom, E.; Welling, M. E (n) equivariant graph neural networks. International conference on machine learning. 2021; pp 9323–9332
  27. Kumar, K.; Vantassel, J. GNS: A generalizable Graph Neural Network-based simulator for particulate and fluid modeling. 2022,
  28. Hu, W.; Shuaibi, M.; Das, A.; Goyal, S.; Sriram, A.; Leskovec, J.; Parikh, D.; Zitnick, C. L. ForceNet: A Graph Neural Network for Large-Scale Quantum Calculations. 2021
  29. Raghunathan Ramakrishnan, M. R., Pavlo O. Dral; von Lilienfeld, O. A. Quantum chemistry structures and properties of 134 kilo molecules. Scientific Data 2014,
  30. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. 2019
  31. Xie, T.; Fu, X.; Ganea, O.-E.; Barzilay, R.; Jaakkola, T. S. Crystal Diffusion Variational Autoencoder for Periodic Material Generation. International Conference on Learning Representations. 2022
  32. Arts, M.; Satorras, V. G.; Huang, C.-W.; Zuegner, D.; Federici, M.; Clementi, C.; Noé, F.; Pinsler, R.; Berg, R. v. d. Two for One: Diffusion Models and Force Fields for Coarse-Grained Molecular Dynamics. arXiv preprint arXiv:2302.00600 2023,
  33. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. 2019,
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Zehua Zhang (16 papers)
  2. Zijie Li (14 papers)
  3. Amir Barati Farimani (121 papers)

Summary

  • The paper presents a novel mask pretraining strategy that enhances GNN performance by learning robust molecular representations.
  • The method achieves significant improvements in the accuracy of force and energy predictions while accelerating model convergence.
  • Results demonstrate the approach's broad applicability across GNN architectures, offering promising advances for computational chemistry and materials science.

Improving Graph Neural Networks for Molecular Dynamics through Mask Pretraining

Introduction to Molecular Dynamics and the Role of GNNs

In the field of computational chemistry and materials science, accurately modeling molecular dynamics is essential for a variety of applications, from drug discovery to the design of new materials. Central to these studies is the potential energy surface (PES), which describes the energy of a system with respect to the positions of all atoms involved. Traditionally, methods such as Density Functional Theory (DFT) have been employed to paper PES with high accuracy, albeit at a significant computational cost, especially for large systems. Alternatively, empirical interatomic potentials offer faster computations but lack the accuracy and adaptability seen in ab initio methods.

With advancements in ML, new methods have emerged to approximate PES with near ab initio accuracy at a fraction of the computational cost. Notably, the incorporation of Graph Neural Networks (GNNs), which directly learn from atomic coordinates, represents a significant advance in the modeling of molecular dynamics. Particularly, the development of neural network-based potentials using GNNs has shown promise in accurately representing PES while maintaining computational efficiency.

The Challenge with GNN-based Neural Potentials

Despite their potential, the performance of GNN-based neural potentials heavily depends on the availability of large, accurately labeled datasets, which are expensive and challenging to generate for complex molecular systems. To address this bottleneck, various pretraining strategies have been investigated, aiming to improve the data efficiency and performance of GNNs in molecular modeling tasks. These strategies involve pretraining GNNs on related tasks with ample available data before finetuning them on the target molecular dynamics prediction tasks.

Our Contribution: Mask Pretraining for Enhancing GNN Performance

We propose a novel pretraining strategy specifically designed for improving the performance of GNN-based neural potentials. Our approach involves pretraining GNNs by masking out spatial information of certain atoms—in our case, a hydrogen atom in water molecules—and then predicting the masked information based on the context provided by the unmasked parts of the molecule. This pretraining task forces GNNs to learn meaningful representations of molecular structures and the underlying physical interactions without the need for explicit labeling, thus enhancing the model's ability to predict potential energy and forces with improved accuracy and speed upon subsequent finetuning.

Experimental Validation

Our comprehensive experiments showcase the effectiveness of the proposed pretraining method. When compared to GNNs trained from scratch or those pretrained using other methods such as denoising, our approach consistently delivers superior performance across different metrics. Notably, we observe significant improvements in both the accuracy of force and energy predictions and the speed of model convergence. These results are robust across different GNN architectures, including those focused on force-centric and energy-centric predictions, highlighting the general applicability of our pretraining strategy.

Future Directions and Implications

The findings from our work open up new avenues for further exploration and application of GNNs in the paper of molecular dynamics. By demonstrating the utility of mask pretraining, we not only provide a novel method for enhancing the performance of neural potentials but also underscore the potential of pretraining strategies in overcoming data limitations inherent in modeling complex molecular systems. Future work may extend our approach to larger and more varied datasets or explore the integration of our pretraining strategy with other GNN architectures and learning paradigms.

In conclusion, our paper introduces a promising direction for the use of GNNs in computational chemistry and materials science, potentially accelerating advancements in drug discovery, materials design, and beyond. The ability to accurately model molecular dynamics with improved data efficiency has far-reaching implications, opening the door to new discoveries and innovations across sciences.

X Twitter Logo Streamline Icon: https://streamlinehq.com