Non-Autonomous Time-Consistent Control
- Non-autonomous time-consistent problems are defined in stochastic control with explicit time-dependent parameters that ensure an optimal control policy remains valid over time.
- A decomposition method splits the problem into a finite-horizon time-inconsistent segment and an infinite-horizon tail with exponential discounting solved via classical dynamic programming.
- Equilibrium strategies in this framework reconcile recursive cost structures, utilizing Hamilton-Jacobi-Bellman equations and backward stochastic differential equations for robust solutions.
A non-autonomous time-consistent problem in stochastic optimal control refers to a dynamic decision problem in which system parameters or performance criteria depend explicitly on time, yet the optimal control policy maintains time consistency—meaning that a strategy devised at the initial time remains optimal at all subsequent times under a suitable information structure or after appropriate augmentation of the state. The theory emerges prominently when infinite-horizon stochastic control problems with initially time-inconsistent preferences or constraints are decomposed under regularity or rationality assumptions, leading to a well-posed non-autonomous, time-consistent "tail" problem that can be addressed by dynamic programming.
1. Problem Decomposition and Setting
Infinite-horizon stochastic optimal control problems with general (possibly non-exponential) discounting or cost structures are not typically time consistent: an optimal control planned at time may lose its optimality as time evolves. However, if it is assumed that decision-makers become rational, or the cost functional simplifies (e.g., discounting becomes exponential) after some finite time , then the problem can be rigorously split into two segments:
- Finite-Horizon Time-Inconsistent Segment (): The system operates under general, typically non-exponential, possibly non-stationary discounting or cost structures, causing time-inconsistency.
- Infinite-Horizon Non-Autonomous Time-Consistent Tail (): Beyond a threshold, the system cost reduces to a classical (e.g., exponentially discounted) form, and the usual time-consistent dynamic programming approach is valid.
Formally, this is encapsulated by imposing for :
where is the effective discounting applied beyond [(Wei et al., 18 Sep 2025), (1.15)].
2. Mathematical Formulation of the Non-Autonomous Time-Consistent Problem
On , the control problem is recast as a classical non-autonomous stochastic optimal control problem. The system evolves as a controlled diffusion: where and are continuous in and satisfy uniform Lipschitz and non-degeneracy assumptions [(H1)].
The performance criterion is of Bolza type: with the running cost, typically bounded or for [(H2)].
The value function and associated HJB equation are:
where
The optimal (feedback) law is
The system is non-autonomous due to explicit -dependence in , , and , but time consistency is restored by the exponential discount.
3. Equilibrium Strategy and Matching at the Interface
The overall infinite-horizon problem is handled by constructing an equilibrium strategy:
- The "tail" (non-autonomous, time-consistent) control is derived on via the time-consistent HJB theory.
- On , where time-inconsistency persists, an "equilibrium HJB" or similar local equilibrium methodology is invoked to produce a policy yielding local time-consistency.
The essential mechanism is to ensure that the solution at —denoted —serves as the continuation value for the earlier interval: where describes the pre- discount. The equilibrium strategy on is constructed such that is locally optimal and consistent with the tail control policy (Wei et al., 18 Sep 2025).
4. Recursive Cost Structures
In addition to Bolza-type (additive) costs, the methodology extends to recursive cost problems formulated via backward stochastic differential equations (BSDEs). Consider
$dY(s) = -g\bigl(t, s, X(s), u(s), Y(s), Z(s)\bigr)\, ds + Z(s)\, dW(s), \quad Y(T) = h\bigl(t, X(T)\bigr), \tag{4.1}-\tag{4.2}$
Assuming a recursive cost structure, e.g., , time-consistency in the tail is approached analogously to the Bolza case by matching the solution at with the infinite-horizon recursive value function.
5. Key Assumptions, Verification, and Implications
Key conditions:
- (H1): Regularity and uniform ellipticity/controllability of , .
- (H2): Integrability and growth conditions on ensuring that is finite.
The splitting method relies on the assumption (1.15) (eventual exponential discount), which enables the "tail" problem to inherit all the properties of classical time-consistent stochastic control.
Time consistency is achieved in the tail because with exponential discount and possibly non-stationary, but regular, coefficients, the dynamic programming principle (DPP) holds: the BeLLMan equation yields optimal feedbacks that remain optimal irrespective of the starting time, as long as the augmented state fully characterizes the system.
6. Broader Context and Connections
The decomposition strategy parallels the general paradigm for handling time-inconsistency in stochastic control:
- The tail non-autonomous, time-consistent problem admits classical verification arguments and HJB analysis.
- The initial segment, inherently time-inconsistent, is treated via equilibrium or generalized (game-theoretic or fixed-point) methods.
- The resulting equilibrium strategy "locks in" the solution, precluding incentive to deviate when the tail problem is rejoined.
This approach extends to problems with recursive utilities (e.g., via BSDEs), and the theory is robust to inclusion of running and terminal costs, non-autonomous coefficients, and both degenerate and non-degenerate diffusions.
7. Conclusion
A non-autonomous time-consistent problem, as extracted from the tail of a decomposed infinite-horizon time-inconsistent control problem (Wei et al., 18 Sep 2025), is mathematically formulated as a stochastic optimal control with explicitly time-dependent coefficients and exponential discounting starting after a sufficiently large time. This framework ensures that the dynamic programming principle and time-consistent policies are applicable in the tail region, and serves as the anchor for constructing equilibrium solutions to the original, more general time-inconsistent stochastic control problem.