Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Stacked tensorial neural networks for reduced-order modeling of a parametric partial differential equation (2312.14979v1)

Published 21 Dec 2023 in cs.LG and cs.NE

Abstract: Tensorial neural networks (TNNs) combine the successes of multilinear algebra with those of deep learning to enable extremely efficient reduced-order models of high-dimensional problems. Here, I describe a deep neural network architecture that fuses multiple TNNs into a larger network, intended to solve a broader class of problems than a single TNN. I evaluate this architecture, referred to as a "stacked tensorial neural network" (STNN), on a parametric PDE with three independent variables and three parameters. The three parameters correspond to one PDE coefficient and two quantities describing the domain geometry. The STNN provides an accurate reduced-order description of the solution manifold over a wide range of parameters. There is also evidence of meaningful generalization to parameter values outside its training data. Finally, while the STNN architecture is relatively simple and problem agnostic, it can be regularized to incorporate problem-specific features like symmetries and physical modeling assumptions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. B. Haasdonk, Model reduction and approximation: theory and algorithms 15, 65 (2017).
  2. A. Miranville and S. Zelik, “Chapter 3 attractors for dissipative partial differential equations in bounded and unbounded domains,” in Handbook of Differential Equations: Evolutionary Equations (Elsevier, 2008) p. 103–200.
  3. J. C. Robinson, Cambridge texts in applied mathematics: Infinite-dimensional dynamical systems: An introduction to dissipative parabolic PDEs and the theory of global attractors series number 28, Cambridge texts in applied mathematics (Cambridge University Press, Cambridge, England, 2001).
  4. S. A. Faroughi, N. Pawar, C. Fernandes, M. Raissi, S. Das, N. K. Kalantari,  and S. K. Mahjour, “Physics-guided, physics-informed, and physics-encoded neural networks in scientific computing,”  (2022).
  5. S. Goswami, A. Bora, Y. Yu,  and G. E. Karniadakis, “Physics-informed deep neural operator networks,” in Machine Learning in Modeling and Simulation: Methods and Applications, edited by T. Rabczuk and K.-J. Bathe (Springer International Publishing, Cham, 2023) pp. 219–254.
  6. S. Fresca and A. Manzoni, Comput. Methods. Appl. Mech. Eng. 388, 114181 (2022).
  7. C. Cercignani, Mathematical Methods in Kinetic Theory (Springer US, 1990).
  8. G. C. Pomraning, Nucl. Sci. Eng. 112, 239–255 (1992).
  9. R. G. McClarren, Trans. Theory Stat. Phys. 39, 73–109 (2010).
  10. A. Kokhanovsky, ed., Springer Series in Light Scattering: Volume 3: Radiative Transfer and Light Scattering (Springer International Publishing, 2019).
  11. D. Mumford, “Elastica and computer vision,” in Algebraic Geometry and its Applications, edited by C. L. Bajaj (Springer New York, New York, NY, 1994) pp. 491–506.
  12. C. G. Wagner and R. Beals, J. Phys. A 52, 115204 (2019).
  13. D. Zwillinger and V. Dobrushkin, Handbook of differential equations, 4th ed., Advances in Applied Mathematics (Chapman & Hall/CRC, Philadelphia, PA, 2021).
  14. “CuPy: Numpy & scipy for GPU (version 10.6.0),” https://github.com/cupy/cupy.
  15. J. A. Reyes and E. M. Stoudenmire, Mach. Learn. Sci. Technol. 2, 035036 (2021).
  16. T. G. Kolda and B. W. Bader, SIAM Review 51, 455–500 (2009).
  17. I. V. Oseledets, SIAM J. Sci. Comput. 33, 2295–2317 (2011).
  18. S. L. Brunton and J. N. Kutz, Data-Driven Science and Engineering (Cambridge University Press, 2019).
  19. K. Kormann, SIAM J. Sci. Comput. 37, B613–B632 (2015).
  20. M. Kiffner and D. Jaksch, Phys. Rev. Fluids 8, 124101 (2023).
  21. E. Ye and N. Loureiro, arXiv preprint arXiv:2311.07756  (2023).
  22. M. E. Widder and U. M. Titulaer, J. Stat. Phys. 56, 471–498 (1989).
  23. L. A. Viehland, “Moment methods for solving the boltzmann equation,” in Gaseous Ion Mobility, Diffusion, and Reaction (Springer International Publishing, Cham, 2018) pp. 127–154.
  24. M. Abadi et al., “TensorFlow: Large-scale machine learning on heterogeneous systems,”  (2015), software available from tensorflow.org.
  25. “t3f (version 1.1.0),” https://github.com/Bihaqo/t3f.
  26. X. Glorot and Y. Bengio, in Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Proc. Mach. Learn. Res., Vol. 9, edited by Y. W. Teh and M. Titterington (PMLR, 2010) pp. 249–256.
  27. C. G. Wagner, Mathematics of nonequilibrium steady states in dilute active matter, Ph.D. thesis (2020).
  28. P. Morse and H. Feshbach, Methods of Theoretical Physics (McGraw-Hill, 1953).
  29. N. Papamichael and N. Stylianopoulos, Numerical conformal mapping: Domain decomposition and the mapping of quadrilaterals (World Scientific Publishing, Singapore, Singapore, 2010).
  30. J. Berg and K. Nyström, Neurocomputing 317, 28–41 (2018).
  31. J.-K. Seo, Scientific Reports 12 (2022), 10.1038/s41598-022-18315-4.

Summary

We haven't generated a summary for this paper yet.