Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
166 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Inverse Optimal Control for Linear Quadratic Tracking with Unknown Target States (2402.17247v1)

Published 27 Feb 2024 in eess.SY, cs.SY, and math.OC

Abstract: This paper addresses the inverse optimal control for the linear quadratic tracking problem with a fixed but unknown target state, which aims to estimate the possible triplets comprising the target state, the state weight matrix, and the input weight matrix from observed optimal control input and the corresponding state trajectories. Sufficient conditions have been provided for the unique determination of both the linear quadratic cost function as well as the target state. A computationally efficient and numerically reliable parameter identification algorithm is proposed by equating optimal control strategies with a system of linear equations, and the associated relative error upper bound is derived in terms of data volume and signal-to-noise ratio. Moreover, the proposed inverse optimal control algorithm is applied for the joint cluster coordination and intent identification of a multi-agent system. By incorporating the structural constraint of the Laplace matrix, the relative error upper bound can be reduced accordingly. Finally, the algorithm's efficiency and accuracy are validated by a vehicle-on-a-lever example and a multi-agent formation control example.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Mosek ApS. Mosek optimization toolbox for matlab. User’s Guide and Reference Manual, Version, 4:1, 2019.
  2. Trust-region inverse reinforcement learning. IEEE Transactions on Automatic Control, 69(2):1037–1044, 2024.
  3. Online inverse optimal control for time-varying cost weights. Biomimetics, 9(2):84, 2024.
  4. Sequential inverse optimal control of discrete-time systems. IEEE/CAA Journal of Automatica Sinica, 11(3):1–14, 2024.
  5. A low complexity approach to model-free stochastic inverse linear quadratic control. IEEE Access, 10:9298–9308, 2022.
  6. Inverse optimal control from incomplete trajectory observations. The International Journal of Robotics Research, 40(6-7):848–865, 2021.
  7. Distributed inverse optimal control. Automatica, 129:109658, 2021.
  8. Learning from sparse demonstrations. IEEE Transactions on Robotics, 39(1):645–664, 2023.
  9. Learning from human directional corrections. IEEE Transactions on Robotics, 39(1):625–644, 2023.
  10. Rudolf Emil Kalman. When is a linear control system optimal? 1964.
  11. Inverse stochastic optimal control for linear-quadratic tracking. In 2023 42nd Chinese Control Conference (CCC), pages 1430–1435. IEEE, 2023.
  12. Inverse kalman filtering for systems with correlated noises. In 2023 62nd IEEE Conference on Decision and Control (CDC), pages 3626–3631, 2023.
  13. Inverse reinforcement learning for multi-player noncooperative apprentice games. Automatica, 145:110524, 2022.
  14. Inverse q-learning using input–output data. IEEE Transactions on Cybernetics, 54(2):728–738, 2024.
  15. Off-policy inverse q-learning for discrete-time antagonistic unknown systems. Automatica, 155:111171, 2023.
  16. Composite adaptive online inverse optimal control approach to human behavior learning. Information Sciences, 638:118977, 2023.
  17. An extended inverse reinforcement learning control for target-expert systems with different parameters. In 2023 IEEE International Symposium on Product Compliance Engineering - Asia (ISPCE-ASIA), pages 1–6, 2023.
  18. J. Löfberg. Yalmip : A toolbox for modeling and optimization in matlab. In In Proceedings of the CACSD Conference, Taipei, Taiwan, 2004.
  19. Matrix differential calculus with applications in statistics and econometrics. John Wiley & Sons, 2019.
  20. Constrained inverse optimal control with application to a human manipulation task. IEEE Transactions on Control Systems Technology, 29(2):826–834, 2019.
  21. Inverse optimal control and inverse noncooperative dynamic game theory. Springer, 2022.
  22. Finite-horizon inverse optimal control for discrete-time nonlinear systems. Automatica, 87:442–446, 2018.
  23. Online inverse optimal control for control-constrained discrete-time systems on finite and infinite horizons. Automatica, 120:109109, 2020.
  24. A convex approach to inverse optimal control and its application to modeling human locomotion. In 2012 IEEE International Conference on Robotics and Automation, pages 531–536. IEEE, 2012.
  25. Optimality and identification of dynamic models in systems biology: an inverse optimal control framework. Bioinformatics, 34(14):2433–2440, 2018.
  26. Inverse optimal control with time-varying objectives: application to human jumping movement analysis. Scientific reports, 10(1):11174, 2020.
  27. Huai-Ning Wu and Mi Wang. Distributed adaptive inverse differential game approach to leader’s behavior learning for multiple autonomous followers. IEEE Transactions on Artificial Intelligence, 4(6):1666–1678, 2023.
  28. Huai-Ning Wu and Mi Wang. Human-in-the-loop behavior modeling via an integral concurrent adaptive inverse reinforcement learning. IEEE Transactions on Neural Networks and Learning Systems, pages 1–12, 2023.
  29. Huai-Ning Wu and Mi Wang. Learning human behavior in shared control: Adaptive inverse differential game approach. IEEE Transactions on Cybernetics, pages 1–11, 2023.
  30. Inverse reinforcement learning in tracking control based on inverse optimal control. IEEE Transactions on Cybernetics, 52(10):10570–10581, 2022.
  31. Inverse reinforcement q-learning through expert imitation for discrete-time systems. IEEE Transactions on Neural Networks and Learning Systems, 34(5):2386–2399, 2023.
  32. Optimization of formation for multi-agent systems based on lqr. Frontiers of Information Technology & Electronic Engineering, 17(2):96–109, 2016.
  33. System identification approach for inverse optimal control of finite-horizon linear quadratic regulators. Automatica, 129:109636, 2021.
  34. Inverse linear quadratic dynamic games using partial state observations. Automatica, 145:110534, 2022.
  35. Statistically consistent inverse optimal control for discrete-time indefinite linear-quadratic systems. arXiv preprint arXiv:2212.08426, 2022.
  36. Inverse linear-quadratic discrete-time finite-horizon optimal control for indistinguishable homogeneous agents: A convex optimization approach. Automatica, 148:110758, 2023.
  37. Inverse optimal control for averaged cost per stage linear quadratic regulators. Systems & Control Letters, 183:105658, 2024.
  38. Statistically consistent inverse optimal control for linear-quadratic tracking with random time horizon. In 2022 41st Chinese Control Conference (CCC), pages 1515–1522, 2022.
  39. Online inverse identification of noncooperative dynamic games. In 2021 IEEE International Conference on Unmanned Systems (ICUS), pages 408–413, 2021.
  40. Mo Zhou. Valuing environmental amenities through inverse optimization: Theory and case study. Journal of Environmental Economics and Management, 83:217–230, 2017.
Citations (1)

Summary

We haven't generated a summary for this paper yet.