Koopman Invertible Autoencoder: Leveraging Forward and Backward Dynamics for Temporal Modeling (2309.10291v1)
Abstract: Accurate long-term predictions are the foundations for many machine learning applications and decision-making processes. However, building accurate long-term prediction models remains challenging due to the limitations of existing temporal models like recurrent neural networks (RNNs), as they capture only the statistical connections in the training data and may fail to learn the underlying dynamics of the target system. To tackle this challenge, we propose a novel machine learning model based on Koopman operator theory, which we call Koopman Invertible Autoencoders (KIA), that captures the inherent characteristic of the system by modeling both forward and backward dynamics in the infinite-dimensional Hilbert space. This enables us to efficiently learn low-dimensional representations, resulting in more accurate predictions of long-term system behavior. Moreover, our method's invertibility design guarantees reversibility and consistency in both forward and inverse operations. We illustrate the utility of KIA on pendulum and climate datasets, demonstrating 300% improvements in long-term prediction capability for pendulum while maintaining robustness against noise. Additionally, our method excels in long-term climate prediction, further validating our method's effectiveness.
- Forecasting sequential data using consistent koopman autoencoders. In International Conference on Machine Learning, pages 475–485. PMLR, 2020.
- Consistent dynamic mode decomposition. SIAM Journal on Applied Dynamical Systems, 18(3):1565–1585, 2019.
- An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint arXiv:1803.01271, 2018.
- Physics guided neural networks for spatio-temporal super-resolution of turbulent flows. In Uncertainty in Artificial Intelligence, pages 118–128. PMLR, 2022.
- Learning long-term dependencies with gradient descent is difficult. IEEE transactions on neural networks, 5(2):157–166, 1994.
- Multifactor sequential disentanglement via structured koopman autoencoders. In The Eleventh International Conference on Learning Representations, 2023.
- Quasi-recurrent neural networks. arXiv preprint arXiv:1611.01576, 2016.
- Data-driven science and engineering: Machine learning, dynamical systems, and control. Cambridge University Press, 2022.
- Opportunities and challenges for machine learning in weather and climate modelling: hard, medium and soft ai. Philosophical Transactions of the Royal Society A, 379(2194):20200083, 2021.
- Gate-variants of gated recurrent unit (gru) neural networks. In 2017 IEEE 60th international midwest symposium on circuits and systems (MWSCAS), pages 1597–1600. IEEE, 2017.
- Nice: Non-linear independent components estimation. arXiv preprint arXiv:1410.8516, 2014.
- Density estimation using real nvp. arXiv preprint arXiv:1605.08803, 2016.
- Sub-seasonal climate forecasting via machine learning: Challenges, analysis, and advances. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 169–177, 2021.
- Long short-term memory. Neural Computation, 9(8):1735–1780, 1997.
- A new sea surface temperature and sea ice boundary dataset for the community atmosphere model. Journal of Climate, 21(19):5145–5153, 2008.
- Physics guided rnns for modeling dynamical systems: A case study in simulating lake temperature profiles. In Proceedings of the 2019 SIAM International Conference on Data Mining, pages 558–566. SIAM, 2019.
- Representation of linguistic form and function in recurrent neural networks. Computational Linguistics, 43(4):761–780, 2017.
- A comparative study on transformer vs rnn in speech applications. In 2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), pages 449–456. IEEE, 2019.
- Non-normal recurrent neural network (nnRNN): Learning long time dependencies while improving expressivity with transient dynamics. In Advances in Neural Information Processing Systems, pages 13591–13601, 2019.
- Glow: Generative flow with invertible 1x1 convolutions. Advances in neural information processing systems, 31, 2018.
- Bernard O Koopman. Hamiltonian systems and transformation in hilbert space. Proceedings of the National Academy of Sciences, 17(5):315–318, 1931.
- Modeling long-and short-term temporal patterns with deep neural networks. In The 41st international ACM SIGIR conference on research & development in information retrieval, pages 95–104, 2018.
- Linearization in the large of nonlinear systems and koopman operator spectrum. Physica D: Nonlinear Phenomena, 242(1):42–53, 2013.
- Cheap orthogonal constraints in neural networks: A simple parametrization of the orthogonal and unitary group. In International Conference on Machine Learning, pages 3794–3803. PMLR, 2019.
- Time-series forecasting with deep learning: a survey. Philosophical Transactions of the Royal Society A, 379(2194):20200209, 2021.
- Learning to encode position for transformer with continuous dynamical model. In International conference on machine learning, pages 6327–6335. PMLR, 2020.
- Deep learning for universal linear embeddings of nonlinear dynamics. Nature Communications, 9(1):4950, 2018.
- Vampnets for deep learning of molecular kinetics. Nature communications, 9(1):1–11, 2018.
- Koopman operator in systems and control. Springer, 2020.
- Machine learning: a practical approach on the statistical learning theory. Springer, 2018.
- Igor Mezić. Analysis of fluid flows via spectral properties of the koopman operator. Annual Review of Fluid Mechanics, 45:357–378, 2013.
- On the difficulty of training recurrent neural networks. In International Conference on Machine Learning, pages 1310–1318, 2013.
- Process-guided deep learning predictions of lake water temperature. Water Resources Research, 55(11):9173–9190, 2019.
- Analyzing koopman approaches to physics-informed machine learning for long-term sea-surface temperature forecasting. arXiv preprint arXiv:2010.00399, 2020.
- Peter J Schmid. Dynamic mode decomposition of numerical and experimental data. Journal of Fluid Mechanics, 656:5–28, 2010.
- Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing, 45(11):2673–2681, 1997.
- The performance of lstm and bilstm in forecasting time series. In 2019 IEEE International Conference on Big Data (Big Data), pages 3285–3292. IEEE, 2019.
- Learning Koopman invariant subspaces for dynamic mode decomposition. In Advances in Neural Information Processing Systems, pages 1130–1140, 2017.
- Invertibility aware integration of static and time-series data: An application to lake temperature modeling. In Proceedings of the 2022 SIAM International Conference on Data Mining (SDM), pages 702–710. SIAM, 2022.
- Koopman neural operator forecaster for time-series with temporal distributional shifts. In The Eleventh International Conference on Learning Representations, 2023.
- Towards physics-informed deep learning for turbulent flow prediction. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1457–1466, 2020.
- Integrating scientific knowledge with machine learning for engineering and environmental systems. ACM Computing Surveys, 55(4):1–37, 2022.
- A data–driven approximation of the koopman operator: Extending dynamic mode decomposition. Journal of Nonlinear Science, 25:1307–1346, 2015.
- Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, pages 38–45, 2020.
- A review of recurrent neural networks: Lstm cells and network architectures. Neural computation, 31(7):1235–1270, 2019.
- Object-aware aggregation with bidirectional temporal graph for video captioning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8327–8336, 2019.