Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Hybrid$^2$ Neural ODE Causal Modeling and an Application to Glycemic Response (2402.17233v2)

Published 27 Feb 2024 in cs.LG, stat.AP, and stat.ME

Abstract: Hybrid models composing mechanistic ODE-based dynamics with flexible and expressive neural network components have grown rapidly in popularity, especially in scientific domains where such ODE-based modeling offers important interpretability and validated causal grounding (e.g., for counterfactual reasoning). The incorporation of mechanistic models also provides inductive bias in standard blackbox modeling approaches, critical when learning from small datasets or partially observed, complex systems. Unfortunately, as the hybrid models become more flexible, the causal grounding provided by the mechanistic model can quickly be lost. We address this problem by leveraging another common source of domain knowledge: \emph{ranking} of treatment effects for a set of interventions, even if the precise treatment effect is unknown. We encode this information in a \emph{causal loss} that we combine with the standard predictive loss to arrive at a \emph{hybrid loss} that biases our learning towards causally valid hybrid models. We demonstrate our ability to achieve a win-win, state-of-the-art predictive performance \emph{and} causal validity, in the challenging task of modeling glucose dynamics post-exercise in individuals with type 1 diabetes.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (76)
  1. Graph network simulators can learn discontinuous, rigid contact dynamics. In Conference on Robot Learning, pages 1157–1167. PMLR, 2023.
  2. Neural ordinary differential equation control of dynamics on graphs. Physical Review Research, 4(1):013221, 2022.
  3. Learning dynamical systems from partial observations. arXiv preprint arXiv:1902.11136, 2019.
  4. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271, 2018.
  5. Mini-review: Glucagon responses in type 1 diabetes–a matter of complexity. Physiological Reports, 9(16):e15009, 2021.
  6. Enhancing the inductive biases of graph neural ode for modeling dynamical systems. arXiv preprint arXiv:2209.10740, 2022.
  7. Combining data assimilation and machine learning to infer unresolved scale parametrization. Philosophical Transactions of the Royal Society A, 379(2194):20200086, 2021.
  8. Barriers to physical activity among patients with type 1 diabetes. Diabetes care, 31(11):2108–2109, 2008.
  9. Viral kinetic modeling: state of the art. Journal of pharmacokinetics and pharmacodynamics, 41:431–443, 2014.
  10. Neural ordinary differential equations. Advances in neural information processing systems, 31, 2018.
  11. Autodifferentiable ensemble kalman filters. SIAM Journal on Mathematics of Data Science, 4(2):801–833, 2022.
  12. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555, 2014.
  13. Physical activity into the meal glucose—insulin model of type 1 diabetes: In silico studies, 2009.
  14. New model of glucose-insulin regulation characterizes effects of physical activity and facilitates personalized treatment evaluation in children and adults with type 1 diabetes. PLOS Computational Biology, 19(2):e1010289, 2023.
  15. Enhancing performance prediction robustness by combining analytical modeling and machine learning. In Proceedings of the 6th ACM/SPEC international conference on performance engineering, pages 145–156, 2015.
  16. Techniques for interpretable machine learning. Communications of the ACM, 63(1):68–77, 2019.
  17. Efficiently modeling long sequences with structured state spaces. arXiv preprint arXiv:2111.00396, 2021.
  18. On the parameterization and initialization of diagonal state space models. Advances in Neural Information Processing Systems, 35:35971–35983, 2022.
  19. Neural closure models for dynamical systems. Proceedings of the Royal Society A, 477(2252):20201004, 2021.
  20. Stable architectures for deep neural networks. Inverse problems, 34(1):014004, 2017.
  21. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  22. Seir modeling of the covid-19 and its dynamics. Nonlinear dynamics, 101:1667–1680, 2020.
  23. DW Hilgemann and Denis Noble. Excitation-contraction coupling and extracellular calcium transients in rabbit atrium: reconstruction of basic cellular mechanisms. Proceedings of the Royal society of London. Series B. Biological sciences, 230(1259):163–205, 1987.
  24. Improving glucose prediction accuracy in physically active adolescents with type 1 diabetes. Journal of diabetes science and technology, 13(4):718–727, 2019.
  25. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
  26. A quantitative description of membrane current and its application to conduction and excitation in nerve. The Journal of physiology, 117(4):500, 1952.
  27. Compartment modeling. Advanced Drug Delivery Reviews, 48(2-3):249–264, 2001.
  28. Neural pharmacodynamic state space modeling. In International Conference on Machine Learning, pages 4500–4510. PMLR, 2021.
  29. Multivariate time series forecasting with dynamic graph neural odes. IEEE Transactions on Knowledge and Data Engineering, 2022.
  30. Physics-informed machine learning. Nature Reviews Physics, 3(6):422–440, 2021.
  31. Patrick Kidger. On neural differential equations, 2022.
  32. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  33. Inferring and validating mechanistic models of neural microcircuits based on spike-train data. Nature communications, 10(1):4933, 2019.
  34. Temporal convolutional networks: A unified approach to action segmentation. In Computer Vision–ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8-10 and 15-16, 2016, Proceedings, Part III 14, pages 47–54. Springer, 2016.
  35. A framework for machine learning of model error in dynamical systems. Communications of the American Mathematical Society, 2(07):283–344, 2022.
  36. Graph neural networks accelerated molecular dynamics. The Journal of Chemical Physics, 156(14), 2022.
  37. Enhancing blood glucose prediction with meal absorption and physical exercise information. arXiv preprint arXiv:1901.07467, 2018.
  38. Neural-ode for pharmacokinetics modeling and its advantage to alternative machine learning models in predicting new dosing regimens. Iscience, 24(7), 2021.
  39. The uva/padova type 1 diabetes simulator: new features. Journal of diabetes science and technology, 8(1):26–34, 2014.
  40. Nonlinear interactions in renal blood flow regulation. American Journal of Physiology-Regulatory, Integrative and Comparative Physiology, 288(5):R1143–R1159, 2005.
  41. Essentials of exercise physiology. Lippincott Williams & Wilkins, 2006.
  42. Learning insulin-glucose dynamics in the wild. In Machine Learning for Healthcare Conference, pages 172–197. PMLR, 2020.
  43. A review of personalized blood glucose prediction strategies for t1dm patients. International journal for numerical methods in biomedical engineering, 33(6):e2833, 2017.
  44. Chemnode: A neural ordinary differential equations framework for efficient chemical kinetic solvers. Energy and AI, 7:100118, 2022.
  45. Human in silico drug trials demonstrate higher accuracy than animal models in predicting clinical pro-arrhythmic cardiotoxicity. Frontiers in physiology, 8:668, 2017.
  46. Hybrid forecasting of chaotic processes: Using machine learning in conjunction with a knowledge-based model. Chaos: An Interdisciplinary Journal of Nonlinear Science, 28(4):041101, 2018.
  47. Judea Pearl. Graphs, causality, and structural equation models. Sociological Methods & Research, 27(2):226–284, 1998.
  48. Hiv-1 dynamics in vivo: virion clearance rate, infected cell life-span, and viral generation time. Science, 271(5255):1582–1586, 1996.
  49. Learning mesh-based simulation with graph networks. arXiv preprint arXiv:2010.03409, 2020.
  50. Graph neural ordinary differential equations. arXiv preprint arXiv:1911.07532, 2019.
  51. Integrating expert odes into neural odes: pharmacology and disease progression. Advances in Neural Information Processing Systems, 34:11364–11383, 2021.
  52. Universal differential equations for scientific machine learning. arXiv preprint arXiv:2001.04385, 2020.
  53. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational physics, 378:686–707, 2019.
  54. Model selection of chaotic systems from data with hidden variables using sparse data assimilation. Chaos: An Interdisciplinary Journal of Nonlinear Science, 32(6), June 2022. ISSN 1089-7682. doi: 10.1063/5.0066066. URL http://dx.doi.org/10.1063/5.0066066.
  55. Continuous-time nonlinear signal processing: a neural network based approach for gray box identification. In Proceedings of IEEE Workshop on Neural Networks for Signal Processing, pages 596–605. IEEE, 1994.
  56. Discrete-vs. continuous-time nonlinear signal processing of cu electrodissolution data. Chemical Engineering Communications, 118(1):25–48, 1992.
  57. Exercise management in type 1 diabetes: a consensus statement. The lancet Diabetes & endocrinology, 5(5):377–390, 2017.
  58. Examining the acute glycemic effects of different types of structured exercise sessions in type 1 diabetes in a real-world setting: The type 1 diabetes and exercise initiative (t1dexi). Diabetes care, 46(4):704–713, 2023.
  59. Graph networks as learnable physics engines for inference and control. In International Conference on Machine Learning, pages 4470–4479. PMLR, 2018.
  60. Learning to simulate complex physics with graph networks. In International conference on machine learning, pages 8459–8468. PMLR, 2020.
  61. Sequence modeling with multiresolution convolutional memory. In International Conference on Machine Learning, pages 31312–31327. PMLR, 2023.
  62. Real-time electronic health record mortality prediction during the covid-19 pandemic: a prospective cohort study. Journal of the American Medical Informatics Association, 28(11):2354–2365, 2021.
  63. Ispad clinical practice consensus guidelines 2022: diabetes technologies: glucose monitoring. Pediatric Diabetes, 23(8):1390–1405, 2022.
  64. A model for human ventricular tissue. American Journal of Physiology-Heart and Circulatory Physiology, 286(4):H1573–H1589, 2004.
  65. Quantifying the impact of physical activity on future glucose trends using machine learning. Iscience, 25(3):103888, 2022.
  66. Causalgnn: Causal-based graph neural networks for spatio-temporal epidemic forecasting. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pages 12191–12199, 2022.
  67. Ee Weinan. A proposal on machine learning via dynamical systems. Communications in Mathematics and Statistics, 1(5):1–11, 2017.
  68. Inflammation, stress, and diabetes. The Journal of clinical investigation, 115(5):1111–1119, 2005.
  69. Integrating scientific knowledge with machine learning for engineering and environmental systems. ACM Computing Surveys, 55(4):1–37, 2022.
  70. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in Neural Information Processing Systems, 34:22419–22430, 2021.
  71. Learning large-scale subsurface simulations with a hybrid graph network simulator. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 4184–4194, 2022.
  72. The causal-neural connection: Expressiveness, learnability, and inference. Advances in Neural Information Processing Systems, 34:10823–10836, 2021.
  73. Neural causal models for counterfactual identification and estimation. arXiv preprint arXiv:2210.00035, 2022.
  74. Benchmarking machine learning algorithms on blood glucose prediction for type i diabetes in comparison with classical time-series models. IEEE Transactions on Biomedical Engineering, 67(11):3101–3124, 2020.
  75. Systems biology informed deep learning for inferring parameters and hidden dynamics. PLoS computational biology, 16(11):e1007575, 2020.
  76. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pages 11106–11115, 2021.
Citations (2)

Summary

We haven't generated a summary for this paper yet.