Learning linear dynamical systems under convex constraints (2303.15121v3)
Abstract: We consider the problem of finite-time identification of linear dynamical systems from $T$ samples of a single trajectory. Recent results have predominantly focused on the setup where no structural assumption is made on the system matrix $A* \in \mathbb{R}{n \times n}$, and have consequently analyzed the ordinary least squares (OLS) estimator in detail. We assume prior structural information on $A*$ is available, which can be captured in the form of a convex set $\mathcal{K}$ containing $A*$. For the solution of the ensuing constrained least squares estimator, we derive non-asymptotic error bounds in the Frobenius norm that depend on the local size of $\mathcal{K}$ at $A*$. To illustrate the usefulness of these results, we instantiate them for four examples, namely when (i) $A*$ is sparse and $\mathcal{K}$ is a suitably scaled $\ell_1$ ball; (ii) $\mathcal{K}$ is a subspace; (iii) $\mathcal{K}$ consists of matrices each of which is formed by sampling a bivariate convex function on a uniform $n \times n$ grid (convex regression); (iv) $\mathcal{K}$ consists of matrices each row of which is formed by uniform sampling (with step size $1/T$) of a univariate Lipschitz function. In all these situations, we show that $A*$ can be reliably estimated for values of $T$ much smaller than what is needed for the unconstrained setting.
- Improved algorithms for linear stochastic bandits. In Advances in Neural Information Processing Systems, volume 24, 2011.
- Living on the edge: phase transitions in convex programs with random data. Information and Inference: A Journal of the IMA, 3(3):224–294, 2014.
- Low rank and structured modeling of high-dimensional vector autoregressions. Trans. Sig. Proc., 67(5):1207–1222, 2019.
- Regularized estimation in sparse high-dimensional time series models. The Annals of Statistics, 43(4):1535–1567, 2015.
- Pierre C. Bellec. Sharp oracle inequalities for Least Squares estimators in shape restricted regression. The Annals of Statistics, 46(2):745 – 780, 2018.
- Near-ideal model selection by ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT minimization. The Annals of Statistics, 37(5A):2145 – 2177, 2009.
- Near-optimal signal recovery from random projections: Universal encoding strategies? IEEE Transactions on Information Theory, 52(12):5406–5425, 2006.
- The convex geometry of linear inverse problems. Found. Comput. Math., 12(6):805–849, 2012.
- On risk bounds in isotonic and other shape restricted regression problems. The Annals of Statistics, 43(4):1774 – 1800, 2015.
- Sourav Chatterjee. A new perspective on least squares under convex constraint. The Annals of Statistics, 42(6):2340–2381, 2014.
- Multivariate convex regression at scale. arXiv: 2005.11588, 2021.
- I.C. Demetriou and P. Tzitziris. Infant mortality and economic growth: Modeling by increasing returns and least squares. In Proceedings of the World Congress on Engineering, 2017.
- Sjoerd Dirksen. Tail bounds via generic chaining. Electronic Journal of Probability, 20(none):1 – 29, 2015.
- R.M Dudley. The sizes of compact subsets of Hilbert space and continuity of Gaussian processes. Journal of Functional Analysis, 1(3):290–330, 1967.
- Learning sparse dynamical systems from a single sample trajectory. In 2019 IEEE 58th Conference on Decision and Control (CDC), pages 2682–2689, 2019.
- Y. Gordon. On Milman’s inequality and random subspaces which escape through a mesh in Rnsuperscript𝑅𝑛R^{n}italic_R start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT. In Geometric Aspects of Functional Analysis, pages 84–106, 1988.
- Nonparametric Shape-Restricted Regression. Statistical Science, 33(4):568 – 594, 2018.
- Clifford Hildreth. Point estimates of ordinates of concave functions. Journal of the American Statistical Association, 49(267):598–619, 1954.
- Finite-time identification of stable linear systems optimality of the least-squares estimator. In 2020 59th IEEE Conference on Decision and Control (CDC), pages 996–1001, 2020.
- Oracle inequalities for high dimensional vector autoregressions. Journal of Econometrics, 186(2):325–344, 2015.
- Suprema of chaos processes and the restricted isometry property. Communications on Pure and Applied Mathematics, 67(11):1877–1904, 2014.
- Convex regression in multidimensions: Suboptimality of least squares estimators. arXiv: 2006.02044, 2020.
- High-dimensional regression with noisy and missing data: Provable guarantees with nonconvexity. The Annals of Statistics, 40(3):1637 – 1664, 2012.
- Linear convergence of gradient methods for estimating structured transition matrices in high-dimensional vector autoregressive models. In Advances in Neural Information Processing Systems, volume 34, pages 16751–16763, 2021.
- Estimating structured vector autoregressive models. In Proceedings of The 33rd International Conference on Machine Learning, pages 830–839, 2016.
- Shahar Mendelson. Learning without concentration. In Proceedings of The 27th Conference on Learning Theory, volume 35, pages 25–39, 2014.
- A unified framework for high-dimensional analysis of m-estimators with decomposable regularizers. Statistical Science, 27(4):538–557, 2012.
- Matey Neykov. Gaussian regression with convex constraints. In Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, volume 89, pages 31–38, 2019.
- Shao Qi-Man Pena Victor H., Lai Tze L. Self-Normalized Processes Limit Theory and Statistical Applications. Probability and Its Applications. Springer Berlin Heidelberg, 2009.
- Learning networks of stochastic differential equations. In Advances in Neural Information Processing Systems, volume 23, 2010.
- The generalized Lasso with non-linear observations. IEEE Trans. Inf. Theor., 62(3):1528–1537, 2016.
- On sparse reconstruction from Fourier and Gaussian measurements. Communications on Pure and Applied Mathematics, 61(8):1025–1045, 2008.
- Near optimal finite time identification of arbitrary linear dynamical systems. In Proceedings of the 36th International Conference on Machine Learning, ICML, volume 97, pages 5610–5618, 2019.
- Nonparametric least squares estimation of a multivariate convex regression function. The Annals of Statistics, 39(3):1633 – 1657, 2011.
- Finite time identification in unstable linear systems. Automatica, 96:342–353, 2018.
- Learning without mixing: Towards a sharp analysis of linear system identification. In Proceedings of the 31st Conference On Learning Theory, volume 75, pages 439–473, 2018.
- Large vector auto regressions. arXiv:1106.3915, 2011.
- Mihailo Stojnic. Various thresholds for ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT-optimization in compressed sensing. arXiv:0907.3666, 2009.
- Michel Talagrand. Upper and lower bounds for stochastic processes, volume 60. Springer, 2014.
- Joel A. Tropp. Convex Recovery of a Structured Signal from Independent Random Linear Measurements, pages 67–101. 2015.
- Roman Vershynin. Introduction to the non-asymptotic analysis of random matrices, page 210–268. Cambridge University Press, 2012.
- Roman Vershynin. High-Dimensional Probability: An Introduction with Applications in Data Science. Number 47 in Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, 2018.
- Di Wang and Ruey S. Tsay. Rate-optimal robust estimation of high-dimensional vector autoregressive models. The Annals of Statistics, 51(2):846 – 877, 2023.
- Finite-time analysis of vector autoregressive models under linear restrictions. Biometrika, 108(2):469–489, 2021.