Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Control: Concurrent System Identification and Control Learning with Neural ODE (2401.01836v4)

Published 3 Jan 2024 in cs.AI

Abstract: Controlling continuous-time dynamical systems is generally a two step process: first, identify or model the system dynamics with differential equations, then, minimize the control objectives to achieve optimal control function and optimal state trajectories. However, any inaccuracy in dynamics modeling will lead to sub-optimality in the resulting control function. To address this, we propose a neural ODE based method for controlling unknown dynamical systems, denoted as Neural Control (NC), which combines dynamics identification and optimal control learning using a coupled neural ODE. Through an intriguing interplay between the two neural networks in coupled neural ODE structure, our model concurrently learns system dynamics as well as optimal controls that guides towards target states. Our experiments demonstrate the effectiveness of our model for learning optimal control of unknown dynamical systems. Codes available at https://github.com/chichengmessi/neural_ode_control/tree/main

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. Wiley Online Library. Optimal control applications and methods. Wiley Online Library, 2023.
  2. Pontryagin L Boltyanskii V. Gamkrelidze R. Mishchenko. Mathematical Theory of Optimal Processes. 1961.
  3. E. McShane. The calculus of variations from the beginning through optimal control theory. SIAM J. Control Optim, 26:916–939, 1989.
  4. Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602, 2013.
  5. Richard S Sutton. Integrated architectures for learning, planning, and reacting based on approximating dynamic programming. IEEE Transactions on Neural Networks, 4(5):782–791, 1993.
  6. Ai pontryagin or how artificial neural networks learn to control dynamical systems. Nature Communications, 13(1):333, 2022.
  7. G. E. Karniadakis and et al. Physics-informed machine learning. Nat. Rev. Phys., 3:422–440, 2021.
  8. Pontryagin differentiable programming: an end-to-end learning and control framework. In 33rd International Conference on Advances in Neural Information Processing Systems, NeurIPS, 2020.
  9. Trust region policy optimization. arXiv preprint arXiv:1502.05477, 2015.
  10. Policy gradient methods for reinforcement learning with function approximation. Advances in neural information processing systems, 12(1):1057–1063, 2000.
  11. Actor-critic algorithms. Advances in neural information processing systems, 12(1):1008–1014, 2000.
  12. Model-based reinforcement learning: A survey. 2020.
  13. Planning with diffusion for flexible behavior synthesis. In International Conference on Machine Learning, May 2022. DOI: 10.48550/arXiv.2205.09991, Corpus ID: 248965046.
  14. Data-efficient reinforcement learning with self-predictive representations. In International Conference on Learning Representations (ICLR), 2021.
  15. Recursive time series data augmentation. In International Conference on Learning Representations (ICLR), February 2023. Published: 01 Feb 2023, Last Modified: 13 Feb 2023.
  16. Neural ordinary differential equations. In Advances in Neural Information Processing Systems, pages 6571–6583, 2018.
  17. Neural ordinary differential equations. Advances in neural information processing systems, 31:6571–6583, 2018.
  18. Swagat Kumar. Balancing a cartpole system with reinforcement learning, 2020.
  19. Razvan V. Florian. Correct equations for the dynamics of the cart-pole system. 2005.
  20. Cheng Chi. Theory Embedded Learning. PhD thesis, Mechanical and Industrial Engineering, University of Toronto, Nov 2023.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com