Causal Affine Control Policies
- Causal affine control policies are state-feedback controllers that depend affinely and causally on the current state, enabling robust and constraint-satisfying control synthesis.
- They leverage LMI-based synthesis and initial-state-to-peak gain techniques to enforce quadratic state and input constraints, ensuring performance under uncertainty.
- The approach incorporates a reciprocal change of variables to convexify the design problem, extending its applicability to uncertain, infinite, and receding-horizon formulations with reduced conservatism.
Causal affine control policies are a class of state-feedback controllers for discrete-time linear (possibly time-varying) systems, wherein the control action at each time depends affinely and causally on the current system state. These policies are formulated to enable synthesis of robust, constraint-satisfying feedback for finite, infinite, and receding-horizon control problems subject to convex quadratic state and input constraints. The synthesis leverages robust control techniques, initial-state-to-peak gain arguments for constraint satisfaction, and a convexification strategy via reciprocal change of variables, leading to a semidefinite programming framework that is computationally scalable and less conservative than standard tube-based robust model predictive control approaches (Gramlich et al., 2023).
1. Affine Feedback Policy Parameterization and Causality
Causal affine policies are defined for discrete-time systems of the form
where is the state, is the control input, and is a disturbance (if considered). Policies are restricted to causal affine state-feedback: for each . Causality is immediate: depends only on and fixed matrices precomputed at the beginning of control; no future state or disturbance terms enter into the feedback calculation at .
2. LMI-Based Synthesis for Finite-Horizon, Unconstrained Problems
For finite-horizon optimal control in the absence of uncertainty, control design is posed in terms of synthesizing a sequence of quadratic value functions
and feedback parameters , subject to a decreasing Bellman inequality,
and quadratic state/input constraints of the form
where stage-cost and constraint outputs are affine in and : , . Terminal and initial value function constraints , (with ) are similarly imposed. These synthesis conditions can be written as linear matrix inequalities (LMIs) in , making the problem tractable via convex optimization.
3. Initial-State-to-Peak Gain and Constraint Satisfaction
Constraint satisfaction, especially for all quadratic state and input constraints, is enforced via an initial-state-to-peak gain (energy-to-peak) argument. Specifically, for all in the sublevel set , it is ensured that
by imposing the LMI
This condition guarantees robust satisfaction of all convex quadratic constraints for every trajectory initialized within .
4. Convexification via Reciprocal Change of Variables
The coupled LMIs in are generally non-affine. A reciprocal variable transformation is introduced to render the synthesis problem fully convex. Specifically, set
A slack variable is incorporated to linearize the initial-state bound , where is an initial-state covariance. The transformed LMIs in these reciprocal variables are fully linear and given by:
- Linearized Bellman LMI (cost decrease)
- Linearized initial-state bound
- Linearized terminal bound
- Linearized constraint satisfaction
Once these LMIs are imposed for each and is maximized, the problem reduces to a semidefinite program (SDP) in . The original controller parameters are recovered via .
5. Extensions to Uncertain, Infinite, and Receding-Horizon Formulations
Uncertainty Modeling
Robust synthesis for disturbed systems models disturbances as satisfying quadratic constraints,
The Bellman decrease is replaced by a robust version using the lossless S-procedure, leading to LMIs affine in augmented decision variables, including multipliers for uncertainty.
Infinite-Horizon Control
Infinite-horizon control assumes time-invariant system data for , enforcing stationarity ( for all ). The terminal bound is replaced with a tail inequality at step , yielding a single LMI that certifies infinite-horizon robust performance and constraint satisfaction.
Receding-Horizon (Robust Model Predictive Control)
A receding-horizon scheme solves the infinite-horizon SDP at each stage with shifted data , obtaining stage-specific policies . After applying the first control action, the state is updated and the optimization is repeated at the next time step. This scheme achieves recursive feasibility, robust constraint satisfaction, and asymptotic performance as .
6. Comparison to Tube-Based Robust Model Predictive Control
Tube-based robust MPC employs a nominal trajectory and an offline-designed, time-invariant error tube (e.g., via robust invariant sets). Online optimization is limited to the nominal variables via a quadratic program (QP), with tightened constraints imposed. While recursive feasibility is maintained and computational scaling is in scalar variables, conservatism is typically high. This arises due to lack of online tube optimization and limitations of uncertainty modeling (commonly norm-bounds, not general LFT).
In contrast, the LMI-based MPC synthesis with causal affine policies jointly optimizes the feedback gains online via an SDP, can handle general linear fractional transformation (LFT) uncertainties, and results in significantly less conservatism. Computationally, each step involves an SDP of block-dimension , heavier per iteration than QP but efficient with structure-exploiting solvers (Gramlich et al., 2023). Decision-variable scaling is notably distinct among approaches:
| Approach | Decision Variables | Conservatism |
|---|---|---|
| Tube-MPC | scalars | High |
| Disturbance/Min-Max MPC | Moderate–High | |
| LMI-based Affine Policy | blocks | Significantly lower |
7. Summary and Context
Causal affine control policies parameterized as state-feedback laws enable robust synthesis for constrained linear systems under both deterministic and uncertain dynamics. The key technical innovations—energy-to-peak (initial-state-to-peak) gain for robust constraint satisfaction and convexification via reciprocal transformation—yield a unified, LMI-based synthesis applicable to finite, infinite, and receding-horizon settings. The resulting controllers offer significant advantages in tractability, flexibility with respect to uncertainty descriptions, and worst-case performance, while maintaining computational scalability comparable to, but less conservative than, traditional tube-based MPC approaches (Gramlich et al., 2023).