Data-Driven Iterative Learning Control
- DDILC is a data-centric iterative control strategy that constructs nonparametric Hankel matrix models to capture dynamic system behavior for repetitive tasks.
- It integrates accelerated gradient methods, such as an inexact Nesterov scheme, to handle unknown LTI dynamics, output disturbances, and input constraints.
- A hybrid switching approach between accelerated and classical updates ensures fast initial error reduction with sustained monotonic convergence under structured uncertainties.
Data-Driven Iterative Learning Control (DDILC) is a class of iterative control methods designed for repetitive tasks in dynamical systems, exploiting directly measured input–output data without reliance on explicit parametric system identification. The DDILC approach leverages nonparametric data representations, typically via the Fundamental Lemma of Willems et al., to construct “lifted” models that govern system response across each trial. This methodology robustly addresses unknown LTI system dynamics, output disturbances, and input constraints, and achieves rapid convergence through integrated acceleration strategies. Recent DDILC frameworks also provide analytical convergence guarantees and robust performance under structured disturbances, motivating their application in precision robotics, motion systems, and beyond.
1. Nonparametric System Representation with Hankel Matrices
A central principle of DDILC is the avoidance of conventional model identification in favor of representing system behavior through input–output data trajectories. Following the Fundamental Lemma, an offline data matrix
is used to form a Hankel matrix of depth , where , is the trial length, and is the initialization horizon. Partitioning and boundary block elimination yield : If the input is persistently exciting of order and , every trajectory is captured via
for some . Eliminating yields the lifted SISO trial-domain model
where is block-Toeplitz and depends on the unknown initial state.
Noisy datasets (output disturbance corrupting ) induce rank loss in . The framework applies QR factorization and SVD-based low-rank reconstruction, ensuring a causal, lower-triangular model by projecting onto strictly lower-triangular matrices: resulting in a direct data-driven (DD) system approximation used in subsequent control synthesis.
2. Iterative Learning Control Law and Input Constraints
At each trial , DDILC formulates the tracking problem as a constrained least-squares optimization using the lifted model: with box constraints . Since only inexact gradient information is available due to model approximation and disturbances, the framework employs
A projected gradient update is then applied: where ensures stability. Constraint handling reduces to simple Euclidean projection, making the scheme amenable to systems with box-bounded inputs.
3. Accelerated Iteration: Inexact Nesterov Gradient
To achieve rapid early-stage convergence, the algorithm incorporates Nesterov’s accelerated gradient strategy, adapted for inexact oracle conditions due to data-driven modeling and disturbances. At trial : where . This mirrors the inexact acceleration scheme of Devolder–Glineur–Nesterov (2014), yielding faster reduction of tracking error until approximation errors begin to accumulate.
4. Disturbance Robustness, Oracle Inexactness, and Switching Strategy
Structured (e.g. periodic or bounded) output disturbances propagate into model and gradient errors. Lemma 3.1 yields quantitative bounds,
with as functions of disturbance magnitude , trial length , and model mismatch . The DDILC framework interprets these bounds as a -inexact oracle, ensuring convergence even in the presence of bounded uncertainty.
For optimal performance, the algorithm deploys a hybrid acceleration scheme: accelerated DDILC is run until the error increases over a sliding window , then switches to classical projected-gradient DDILC for guaranteed monotonic error decay.
5. Theoretical Convergence Properties
Convergence under relative degree 1 (invertible ) and bounded disturbance is established:
| Mode | Error Bound | Convergence Rate |
|---|---|---|
| Classical ILC | to neighborhood | |
| Fast ILC (Accelerated) | , but with error accumulation | |
| Hybrid Switching | Explicit upper bound on trials for given accuracy | Combined fast initial, monotonic tail |
Accelerated DDILC achieves rapid initial error reduction, but accumulation of errors eventually dominates, justifying the hybrid switch for sustained monotonic convergence.
6. Empirical Demonstration and Real-World Case Studies
Numerical experiments validate the DDILC approach:
- Toy Example (4th order plant, trials): Direct DD modeling delivers relative modeling error (structured disturbance), significantly outperforming subspace ID ().
- Hybrid ILC Performance: Hybrid DDILC converges as rapidly as fast ILC for the first 50 trials and then outpaces classical ILC, reaching low error with 200 fewer trials.
- Monte Carlo (100 randomized plants): Hybrid DDILC statistically outperforms alternatives for both early-stage ($1$–$50$ trials) and late-stage ($451$–$500$) error norms.
- Precision Robotics Case Study: On a 2-axis high-precision motion platform, the hybrid DDILC achieves sub-micron tracking accuracy within 10 trials, outperforming feedback-only and standard ILC variants, under output disturbance bounds of $0.1$ mm.
7. Advantages, Limitations, and Extensions
Key advantages:
- Eliminates explicit identification steps, operating solely from input–output data via Hankel filtering.
- Robust to structured (periodic) disturbances and systematic model mismatch.
- Achieves fast convergence in early trials, offering rigorous – guarantees.
- Efficient constraint handling via projection.
Limitations:
- Under large unstructured (white) noise, parametric subspace identification may outperform DDILC.
- Sufficient PE (persistency of excitation) and large enough sampled data are required for Hankel construction.
- Hybrid mode/window for switching between accelerated and classical regimes remains heuristic.
Research directions:
- Development of theoretically grounded hybrid switching criteria.
- Extension of the DDILC methodology to MIMO systems and varying reference tasks.
- Integration with and nuclear-norm regularization for noise suppression and data-efficient learning.
- Adaptive tuning of Lipschitz constant , oracle inexactness , and step sizes.
DDILC constitutes an analytically rigorous, empirically validated methodology for repetitive learning control of unknown LTI systems with disturbances and constraints. Its theoretical and practical virtues motivate ongoing efforts for broader applicability, such as extension to general MIMO architectures, adaptive strategies for time-varying environments, and fusion with robust control and regularized machine learning paradigms (Wang et al., 2023).