Automatically Learning Hybrid Digital Twins of Dynamical Systems (2410.23691v1)
Abstract: Digital Twins (DTs) are computational models that simulate the states and temporal dynamics of real-world systems, playing a crucial role in prediction, understanding, and decision-making across diverse domains. However, existing approaches to DTs often struggle to generalize to unseen conditions in data-scarce settings, a crucial requirement for such models. To address these limitations, our work begins by establishing the essential desiderata for effective DTs. Hybrid Digital Twins ($\textbf{HDTwins}$) represent a promising approach to address these requirements, modeling systems using a composition of both mechanistic and neural components. This hybrid architecture simultaneously leverages (partial) domain knowledge and neural network expressiveness to enhance generalization, with its modular design facilitating improved evolvability. While existing hybrid models rely on expert-specified architectures with only parameters optimized on data, $\textit{automatically}$ specifying and optimizing HDTwins remains intractable due to the complex search space and the need for flexible integration of domain priors. To overcome this complexity, we propose an evolutionary algorithm ($\textbf{HDTwinGen}$) that employs LLMs to autonomously propose, evaluate, and optimize HDTwins. Specifically, LLMs iteratively generate novel model specifications, while offline tools are employed to optimize emitted parameters. Correspondingly, proposed models are evaluated and evolved based on targeted feedback, enabling the discovery of increasingly effective hybrid models. Our empirical results reveal that HDTwinGen produces generalizable, sample-efficient, and evolvable models, significantly advancing DTs' efficacy in real-world applications.
- F. Tao, J. Cheng, Q. Qi, M. Zhang, H. Zhang, and F. Sui, “Digital twin-driven product design, manufacturing and service with big data,” The International Journal of Advanced Manufacturing Technology, vol. 94, pp. 3563–3576, 2018.
- J. Corral-Acero, F. Margara, M. Marciniak, C. Rodero, F. Loncaric, Y. Feng, A. Gilbert, J. F. Fernandes, H. A. Bukhari, A. Wajdan et al., “The ‘digital twin’to enable the vision of precision cardiology,” European heart journal, vol. 41, no. 48, pp. 4556–4564, 2020.
- J. Ladyman, J. Lambert, and K. Wiesner, “What is a complex system?” European Journal for Philosophy of Science, vol. 3, pp. 33–67, 2013.
- Q. Qi and F. Tao, “Digital twin and big data towards smart manufacturing and industry 4.0: 360 degree comparison,” Ieee Access, vol. 6, pp. 3585–3593, 2018.
- V. Iranzo and S. Pérez-González, “Epidemiological models and covid-19: a comparative view,” History and Philosophy of the Life Sciences, vol. 43, no. 3, p. 104, 2021.
- I. Bozic, J. G. Reiter, B. Allen, T. Antal, K. Chatterjee, P. Shah, Y. S. Moon, A. Yaqubie, N. Kelly, D. T. Le et al., “Evolutionary dynamics of cancer in response to targeted combination therapy,” elife, vol. 2, p. e00747, 2013.
- R. Rosen, G. Von Wichert, G. Lo, and K. D. Bettenhausen, “About the importance of autonomy and digital twins for the future of manufacturing,” Ifac-papersonline, vol. 48, no. 3, pp. 567–572, 2015.
- T. Erol, A. F. Mendi, and D. Doğan, “The digital twin revolution in healthcare,” in 2020 4th international symposium on multidisciplinary studies and innovative technologies (ISMSIT). IEEE, 2020, pp. 1–7.
- J. R. Koza, “Genetic programming as a means for programming computers by natural selection,” Statistics and computing, vol. 4, pp. 87–112, 1994.
- S. L. Brunton, J. L. Proctor, and J. N. Kutz, “Discovering governing equations from data by sparse identification of nonlinear dynamical systems,” Proceedings of the national academy of sciences, vol. 113, no. 15, pp. 3932–3937, 2016.
- D. Ha and J. Schmidhuber, “Recurrent world models facilitate policy evolution,” Advances in neural information processing systems, vol. 31, 2018.
- J. Yoon, D. Jarrett, and M. Van der Schaar, “Time-series generative adversarial networks,” Advances in neural information processing systems, vol. 32, 2019.
- R. T. Chen, Y. Rubanova, J. Bettencourt, and D. K. Duvenaud, “Neural ordinary differential equations,” Advances in neural information processing systems, vol. 31, 2018.
- M. Raissi, P. Perdikaris, and G. E. Karniadakis, “Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations,” Journal of Computational physics, vol. 378, pp. 686–707, 2019.
- L. Faure, B. Mollet, W. Liebermeister, and J.-L. Faulon, “A neural-mechanistic hybrid approach improving the predictive power of genome-scale metabolic models,” Nature Communications, vol. 14, no. 1, p. 4669, 2023.
- J. Pinto, J. R. Ramos, R. S. Costa, and R. Oliveira, “A general hybrid modeling framework for systems biology applications: Combining mechanistic knowledge with deep neural networks under the sbml standard,” AI, vol. 4, no. 1, pp. 303–318, 2023.
- P. Wang, Z. Zhu, W. Liang, L. Liao, and J. Wan, “Hybrid mechanistic and neural network modeling of nuclear reactors,” Energy, vol. 282, p. 128931, 2023.
- R. Cheng, A. Verma, G. Orosz, S. Chaudhuri, Y. Yue, and J. Burdick, “Control regularization for reduced variance reinforcement learning,” in International Conference on Machine Learning. PMLR, 2019, pp. 1141–1150.
- E. Real, S. Moore, A. Selle, S. Saxena, Y. L. Suematsu, J. Tan, Q. V. Le, and A. Kurakin, “Large-scale evolution of image classifiers,” in International conference on machine learning. PMLR, 2017, pp. 2902–2911.
- T. N. Mundhenk, M. Landajuela, R. Glatt, C. P. Santiago, D. M. Faissol, and B. K. Petersen, “Symbolic regression via neural-guided genetic programming population seeding,” in Proceedings of the 35th International Conference on Neural Information Processing Systems, 2021, pp. 24 912–24 923.
- T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell et al., “Language models are few-shot learners,” Advances in neural information processing systems, vol. 33, pp. 1877–1901, 2020.
- J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia, E. Chi, Q. V. Le, D. Zhou et al., “Chain-of-thought prompting elicits reasoning in large language models,” Advances in neural information processing systems, vol. 35, pp. 24 824–24 837, 2022.
- A. Chowdhery, S. Narang, J. Devlin, M. Bosma, G. Mishra, A. Roberts, P. Barham, H. W. Chung, C. Sutton, S. Gehrmann et al., “Palm: Scaling language modeling with pathways,” Journal of Machine Learning Research, vol. 24, no. 240, pp. 1–113, 2023.
- H. A. Simon, “The architecture of complexity,” Proceedings of the American philosophical society, vol. 106, no. 6, pp. 467–482, 1962.
- T. L. Rogers, B. J. Johnson, and S. B. Munch, “Chaos is not rare in natural ecosystems,” Nature Ecology & Evolution, vol. 6, no. 8, pp. 1105–1111, 2022.
- M. Sokolov, M. von Stosch, H. Narayanan, F. Feidl, and A. Butté, “Hybrid modeling—a key enabler towards realizing digital twins in biopharma?” Current Opinion in Chemical Engineering, vol. 34, p. 100715, 2021.
- S. Chaudhuri, K. Ellis, O. Polozov, R. Singh, A. Solar-Lezama, Y. Yue et al., “Neurosymbolic programming,” Foundations and Trends® in Programming Languages, vol. 7, no. 3, pp. 158–243, 2021.
- A. Tsoularis and J. Wallace, “Analysis of logistic growth models,” Mathematical biosciences, vol. 179, no. 1, pp. 21–55, 2002.
- A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga et al., “Pytorch: An imperative style, high-performance deep learning library,” Advances in neural information processing systems, vol. 32, 2019.
- D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
- J. H. Holland, “Genetic algorithms,” Scientific american, vol. 267, no. 1, pp. 66–73, 1992.
- L. R. Rabiner, “A tutorial on hidden markov models and selected applications in speech recognition,” Proceedings of the IEEE, vol. 77, no. 2, pp. 257–286, 1989.
- R. Kalman, “A new approach to linear filtering and prediction problems,” Trans. ASME, D, vol. 82, pp. 35–44, 1960.
- L. Li, Y. Zhao, D. Jiang, Y. Zhang, F. Wang, I. Gonzalez, E. Valentin, and H. Sahli, “Hybrid deep neural network–hidden markov model (dnn-hmm) based speech emotion recognition,” in 2013 Humaine association conference on affective computing and intelligent interaction. IEEE, 2013, pp. 312–317.
- R. G. Krishnan, U. Shalit, and D. Sontag, “Deep kalman filters,” arXiv preprint arXiv:1511.05121, 2015.
- J. L. Elman, “Finding structure in time,” Cognitive science, vol. 14, no. 2, pp. 179–211, 1990.
- S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural computation, vol. 9, no. 8, pp. 1735–1780, 1997.
- K. Cho, B. Van Merriënboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio, “Learning phrase representations using rnn encoder-decoder for statistical machine translation,” arXiv preprint arXiv:1406.1078, 2014.
- I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to sequence learning with neural networks,” Advances in neural information processing systems, vol. 27, 2014.
- D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation by jointly learning to align and translate,” arXiv preprint arXiv:1409.0473, 2014.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- E. Dupont, A. Doucet, and Y. W. Teh, “Augmented neural odes,” Advances in neural information processing systems, vol. 32, 2019.
- S. I. Holt, Z. Qian, and M. van der Schaar, “Neural laplace: Learning diverse classes of differential equations in the laplace domain,” in International Conference on Machine Learning. PMLR, 2022, pp. 8811–8832.
- M. A. Zaytar and C. El Amrani, “Sequence to sequence weather forecasting with long short-term memory recurrent neural networks,” International Journal of Computer Applications, vol. 143, no. 11, pp. 7–11, 2016.
- J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” arXiv preprint arXiv:1810.04805, 2018.
- L. Sehovac and K. Grolinger, “Deep learning for load forecasting: Sequence to sequence recurrent neural networks with attention,” Ieee Access, vol. 8, pp. 36 411–36 426, 2020.
- S. Holt, A. Hüyük, Z. Qian, H. Sun, and M. van der Schaar, “Neural laplace control for continuous-time delayed systems,” in International Conference on Artificial Intelligence and Statistics. PMLR, 2023, pp. 1747–1778.
- M. Schmidt and H. Lipson, “Distilling free-form natural laws from experimental data,” science, vol. 324, no. 5923, pp. 81–85, 2009.
- Z. Qian, K. Kacprzyk, and M. van der Schaar, “D-CODE: Discovering closed-form ODEs from observed trajectories,” in International Conference on Learning Representations, 2022. [Online]. Available: https://openreview.net/forum?id=wENMvIsxNN
- K. Kacprzyk, Z. Qian, and M. van der Schaar, “D-cipher: discovery of closed-form partial differential equations,” Advances in Neural Information Processing Systems, vol. 36, 2024.
- K. Kacprzyk, T. Liu, and M. van der Schaar, “Towards transparent time series forecasting,” in The Twelfth International Conference on Learning Representations, 2024. [Online]. Available: https://openreview.net/forum?id=TYXtXLYHpR
- S. Cuomo, V. S. Di Cola, F. Giampaolo, G. Rozza, M. Raissi, and F. Piccialli, “Scientific machine learning through physics–informed neural networks: Where we are and what’s next,” Journal of Scientific Computing, vol. 92, no. 3, p. 88, 2022.
- S. Greydanus, M. Dzamba, and J. Yosinski, “Hamiltonian neural networks,” Advances in neural information processing systems, vol. 32, 2019.
- M. Cranmer, S. Greydanus, S. Hoyer, P. Battaglia, D. Spergel, and S. Ho, “Lagrangian neural networks,” arXiv preprint arXiv:2003.04630, 2020.
- Y. Yin, V. Le Guen, J. Dona, E. de Bézenac, I. Ayed, N. Thome, and P. Gallinari, “Augmenting physical models with deep networks for complex dynamics forecasting,” Journal of Statistical Mechanics: Theory and Experiment, vol. 2021, no. 12, p. 124012, 2021.
- N. Takeishi and A. Kalousis, “Physics-integrated variational autoencoders for robust and interpretable generative modeling,” Advances in Neural Information Processing Systems, vol. 34, pp. 14 809–14 821, 2021.
- Z. Qian, W. Zame, L. Fleuren, P. Elbers, and M. van der Schaar, “Integrating expert odes into neural odes: pharmacology and disease progression,” Advances in Neural Information Processing Systems, vol. 34, pp. 11 364–11 383, 2021.
- A. Wehenkel, J. Behrmann, H. Hsu, G. Sapiro, G. Louppe, and J.-H. Jacobsen, “Robust hybrid learning with expert augmentation,” Transactions on Machine Learning Research, 2023. [Online]. Available: https://openreview.net/forum?id=oe4dl4MCGY
- C. Geng, H. Paganetti, and C. Grassberger, “Prediction of Treatment Response for Combined Chemo- and Radiation Therapy for Non-Small Cell Lung Cancer Patients Using a Bio-Mathematical Model,” Scientific Reports, vol. 7, no. 1, p. 13542, Oct. 2017.
- I. Bica, A. M. Alaa, J. Jordon, and M. van der Schaar, “Estimating counterfactual treatment outcomes over time through adversarially balanced representations,” in International Conference on Learning Representations, 2020.
- N. Seedat, F. Imrie, A. Bellot, Z. Qian, and M. van der Schaar, “Continuous-time modeling of counterfactual outcomes using neural controlled differential equations,” arXiv preprint arXiv:2206.08311, 2022.
- V. Melnychuk, D. Frauen, and S. Feuerriegel, “Causal transformer for estimating counterfactual outcomes,” in International Conference on Machine Learning. PMLR, 2022, pp. 15 293–15 329.
- C. C. Kerr, R. M. Stuart, D. Mistry, R. G. Abeysuriya, K. Rosenfeld, G. R. Hart, R. C. Núñez, J. A. Cohen, P. Selvaraj, B. Hagedorn et al., “Covasim: an agent-based model of covid-19 dynamics and interventions,” PLOS Computational Biology, vol. 17, no. 7, p. e1009149, 2021.
- T. Hiltunen, L. Jones, S. Ellner, and N. G. Hairston Jr, “Temporal dynamics of a simple community with intraguild predation: an experimental test,” Ecology, vol. 94, no. 4, pp. 773–779, 2013.
- E. P. Odum and G. W. Barrett, “Fundamentals of ecology,” The Journal of Wildlife Management, vol. 36, no. 4, p. 1372, 1972.
- V. M. M. Alvarez, R. Roşca, and C. G. Fălcuţescu, “Dynode: Neural ordinary differential equations for dynamics modeling in continuous control,” arXiv preprint arXiv:2009.04278, 2020.
- D. E. Rumelhart, G. E. Hinton, and R. J. Williams, “Learning representations by back-propagating errors,” nature, vol. 323, no. 6088, pp. 533–536, 1986.
- S. Holt, A. Hüyük, and M. van der Schaar, “Active observing in continuous-time control,” Advances in Neural Information Processing Systems, vol. 36, 2024.
- S. Holt, Z. Qian, and M. van der Schaar, “Deep generative symbolic regression,” in The Eleventh International Conference on Learning Representations, 2023. [Online]. Available: https://openreview.net/forum?id=o7koEEMA1bR
- K. Kacprzyk, S. Holt, J. Berrevoets, Z. Qian, and M. van der Schaar, “ODE discovery for longitudinal heterogeneous treatment effects inference,” in The Twelfth International Conference on Learning Representations, 2024. [Online]. Available: https://openreview.net/forum?id=pxI5IPeWgW
- T. Liu, N. Astorga, N. Seedat, and M. van der Schaar, “Large language models to enhance bayesian optimization,” in The Twelfth International Conference on Learning Representations, 2024. [Online]. Available: https://openreview.net/forum?id=OOxotBmGol
- C. Yang, X. Wang, Y. Lu, H. Liu, Q. V. Le, D. Zhou, and X. Chen, “Large language models as optimizers,” in The Twelfth International Conference on Learning Representations, 2024. [Online]. Available: https://openreview.net/forum?id=Bb4VGOWELI
- M. Chen, J. Tworek, H. Jun, Q. Yuan, H. P. D. O. Pinto, J. Kaplan, H. Edwards, Y. Burda, N. Joseph, G. Brockman et al., “Evaluating large language models trained on code,” arXiv preprint arXiv:2107.03374, 2021.
- S. Holt, M. R. Luyten, and M. van der Schaar, “L2MAC: Large language model automatic computer for extensive code generation,” in The Twelfth International Conference on Learning Representations, 2024. [Online]. Available: https://openreview.net/forum?id=EhrzQwsV4K
- S. Holt, Z. Qian, T. Liu, J. Weatherall, and M. van der Schaar, “Data-driven discovery of dynamical systems in pharmacology using large language models,” in The Thirty-eighth Annual Conference on Neural Information Processing Systems, 2024.
- N. Astorga, T. Liu, N. Seedat, and M. van der Schaar, “Partially observable cost-aware active-learning with large language models,” in The Thirty-Eighth Annual Conference on Neural Information Processing Systems, 2024.
- W. Bonnaffé and T. Coulson, “Fast fitting of neural ordinary differential equations by bayesian neural gradient matching to infer ecological interactions from time-series data,” Methods in Ecology and Evolution, vol. 14, no. 6, pp. 1543–1563, 2023.
- S. K. Kumar, “On weight initialization in deep neural networks,” arXiv preprint arXiv:1704.08863, 2017.
- A. Graves, S. Fernández, and J. Schmidhuber, “Multi-dimensional recurrent neural networks,” in International conference on artificial neural networks. Springer, 2007, pp. 549–558.
- B. K. Petersen, M. L. Larma, T. N. Mundhenk, C. P. Santiago, S. K. Kim, and J. T. Kim, “Deep symbolic regression: Recovering mathematical expressions from data via risk-seeking policy gradients,” in International Conference on Learning Representations, 2020.
- S. Hsiang, D. Allen, S. Annan-Phan, K. Bell, I. Bolliger, T. Chong, H. Druckenmiller, L. Y. Huang, A. Hultgren, E. Krasovich et al., “The effect of large-scale anti-contagion policies on the covid-19 pandemic,” Nature, vol. 584, no. 7820, pp. 262–267, 2020.
- O. N. Bjørnstad, K. Shea, M. Krzywinski, and N. Altman, “The seirs model for infectious disease dynamics.” Nature methods, vol. 17, no. 6, pp. 557–559, 2020.