Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
113 tokens/sec
GPT-4o
12 tokens/sec
Gemini 2.5 Pro Pro
37 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
3 tokens/sec
DeepSeek R1 via Azure Pro
33 tokens/sec
2000 character limit reached

Dynamical Low-Rank Approximation Strategies for Nonlinear Feedback Control Problems (2501.07439v1)

Published 13 Jan 2025 in math.NA and cs.NA

Abstract: This paper addresses the stabilization of dynamical systems in the infinite horizon optimal control setting using nonlinear feedback control based on State-Dependent Riccati Equations (SDREs). While effective, the practical implementation of such feedback strategies is often constrained by the high dimensionality of state spaces and the computational challenges associated with solving SDREs, particularly in parametric scenarios. To mitigate these limitations, we introduce the Dynamical Low-Rank Approximation (DLRA) methodology, which provides an efficient and accurate framework for addressing high-dimensional feedback control problems. DLRA dynamically constructs a compact, low-dimensional representation that evolves with the problem, enabling the simultaneous resolution of multiple parametric instances in real-time. We propose two novel algorithms to enhance numerical performances: the cascade-Newton-Kleinman method and Riccati-based DLRA (R-DLRA). The cascade-Newton-Kleinman method accelerates convergence by leveraging Riccati solutions from the nearby parameter or time instance, supported by a theoretical sensitivity analysis. R-DLRA integrates Riccati information into the DLRA basis construction to improve the quality of the solution. These approaches are validated through nonlinear one-dimensional and two-dimensional test cases showing transport-like behavior, demonstrating that R-DLRA outperforms standard DLRA and Proper Orthogonal Decomposition-based model order reduction in both speed and accuracy, offering a superior alternative to Full Order Model solutions.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.