Nonparametric Chain Policies: Data-Driven Control
- Nonparametric Chain Policies are data-driven control methods that construct stabilizing feedback controllers from trajectory data without explicit system models.
- They employ nonparametric approximations, such as Gaussian Process and Nyström methods, to linearize the infinite-horizon KL-control problem and compute control actions.
- The framework guarantees practical stability using Recurrent Control Lyapunov Functions and sample complexity bounds, enabling robust control in complex, nonlinear scenarios.
Nonparametric Chain Policies (NCPs) constitute a class of data-driven control synthesis methodologies for nonlinear and stochastic systems. They utilize nonparametric approximation schemes to construct practically stabilizing feedback policies from trajectory data, eschewing classical model-based controller design. Key theoretical results anchor NCPs in Lyapunov-based analysis, nonparametric function approximation, and tractable sample complexity guarantees, establishing them as a rigorous framework for infinite-horizon stabilization and Kullback–Leibler (KL) optimal control in continuous and discrete domains (Siegelmann et al., 5 Oct 2025, Pan et al., 2014).
1. Foundations: Infinite-Horizon KL-Control and Nonparametric Policy Synthesis
NCPs derive from two streams: nonparametric methods for infinite-horizon KL-control in linearly-solvable Markov decision processes (LMDP) (Pan et al., 2014), and certified stabilization of general nonlinear and locally Lipschitz systems (Siegelmann et al., 5 Oct 2025). The control problem typically seeks state-feedback policies minimizing a long-run average cost, often in the form:
with stochastic dynamics .
KL-control models exploit a transformation of the Hamilton–Jacobi–Bellman (HJB) PDE via the desirability function , linearizing the value equation and enabling spectral or sampling-based solution approaches. NCPs broaden this concept, using empirical data (local trajectory segments) and nearest-neighbor assignment to induce closed-loop policies without explicit parametric modeling.
2. Recurrent (Control) Lyapunov Functions and Stability Guarantees
Central to NCPs is the notion of Recurrent Control Lyapunov Functions (R-CLFs), which generalize classical Lyapunov stability by requiring a recurrent, finite-horizon decrease condition. An R-CLF establishes, for each state in a compact region , the existence of a control segment such that
with parameters and bounded by constants through .
This structure allows concatenation of finite-horizon controls, yielding practical exponential stability. Specifically, for a policy constructed via a sequence of such controls, the region can be rendered -practically stable, with the solution trajectory satisfying
where depend explicitly on Lyapunov parameters and system Lipschitz constants (Siegelmann et al., 5 Oct 2025).
3. Nonparametric Chain Policy Construction and the Normalized Nearest-Neighbor Rule
An NCP is specified by a finite "control alphabet" , comprising verified finite-horizon control signals, and an "assignment set" covering the region of interest by balls . At each state , the normalized nearest-neighbor rule assigns the signal where
and concatenates these signals as the state evolves. Policy synthesis proceeds as follows:
- Coverage: Select centers and radii to form a nonparametric cover of the target region.
- Verification: For each , verify finite-horizon decrease via R-CLF conditions ((8a)-(8b) in (Siegelmann et al., 5 Oct 2025)).
- Concatenation: Induce a piecewise policy by chaining assigned signals based on the evolving state and the assignment set K.
This approach does not require explicit system identification; stabilization and convergence are achieved through empirical verification of local controls and geometric covering arguments.
4. Nonparametric Desirability Function Approximation: GP and Nyström Methods
In the context of infinite-horizon KL-control, nonparametric chain policies are realized by approximating the desirability function via Gaussian Processes (GP) or the Nyström method (Pan et al., 2014):
- GP-KL: Treat as a sample from a GP, utilizing kernel regression to fit training data and inferring predictive mean and gradient . The optimal control policy is analytically available:
with online budgeted updates via kernel independence tests and KL-pruning.
- Nyström-KL: Reduce the spectral (eigenfunction) computation to a low-rank approximation using landmark points, solving the eigenproblem on a small matrix and extending to new queries via:
where encodes weighted passive transitions. The control policy is computed by differentiating as before.
Both methods work within the chain policy paradigm, sequentially assigning control segments based on local function evaluations and maintaining tractable computational budgets.
5. Sample Complexity, Covering Arguments, and Incremental Learning
Sample complexity is characterized by explicit bounds for stabilization guarantees. For a region in , desiring exponential convergence rate and precision , there exists a policy with assignment set size
where is a system-dependent constant determined by desired rate and system parameters (see Eq. (11) in (Siegelmann et al., 5 Oct 2025)). The construction proceeds by covering in concentric annuli, each subdivided with a hypercubic grid, and assigning radii and verified controls as per Theorem 3.
Incremental learning is achieved by augmenting the assignment set with new verified triples that satisfy extension conditions, either by direct verification or bootstrapping through an existing center. No re-optimization or recomputation of old data is required—existing controls and certificates remain valid.
6. Connections to Spectral Control, Information Theory, and Nonparametric Learning
NCPs and KL-control policies are theoretically linked via the linearized HJB (Perron–Frobenius eigenproblem) and the free-energy/relative-entropy duality of stochastic control (Fleming–McEneaney, 1995). Gaussian Process and Nyström approximations provide nonparametric analogues to spectral methods (eigenfunction expansions) in parametric frameworks (Pan et al., 2014). This synthesis unites information-theoretic principles, topological entropy concepts (control alphabet minimality), and nonparametric learning, circumventing local minima issues common in parametric optimization.
7. Algorithmic Summary and Practical Considerations
A typical algorithmic workflow for NCPs involves:
- Selection of convergence rate and horizon ,
- Computation of coverage and sample complexity parameters ,
- Empirical or open-loop synthesis of control signals for each grid center ,
- Verification of recurrent decrease conditions,
- Assembly of the assignment set and finalized chain policy ,
- Real-time policy execution by normalized nearest-neighbor assignment, with O() update cost per step in GP/Nyström variants,
- Incremental augmentation by addition of new verified signals as new data and regions of interest arise.
Comparative performance evaluations show that GP-KL yields accurate policies at higher computational expense, while Nyström-KL achieves faster inference with moderate accuracy loss off-manifold (Pan et al., 2014). NCPs support principled tradeoffs between stabilization rate, region enlargement, and sample complexity, and retain feasibility for high-dimensional systems due to their nonparametric, data-driven nature.