Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning linear dynamical systems under convex constraints (2303.15121v3)

Published 27 Mar 2023 in math.ST, cs.SY, eess.SY, math.OC, stat.ML, and stat.TH

Abstract: We consider the problem of finite-time identification of linear dynamical systems from $T$ samples of a single trajectory. Recent results have predominantly focused on the setup where no structural assumption is made on the system matrix $A* \in \mathbb{R}{n \times n}$, and have consequently analyzed the ordinary least squares (OLS) estimator in detail. We assume prior structural information on $A*$ is available, which can be captured in the form of a convex set $\mathcal{K}$ containing $A*$. For the solution of the ensuing constrained least squares estimator, we derive non-asymptotic error bounds in the Frobenius norm that depend on the local size of $\mathcal{K}$ at $A*$. To illustrate the usefulness of these results, we instantiate them for four examples, namely when (i) $A*$ is sparse and $\mathcal{K}$ is a suitably scaled $\ell_1$ ball; (ii) $\mathcal{K}$ is a subspace; (iii) $\mathcal{K}$ consists of matrices each of which is formed by sampling a bivariate convex function on a uniform $n \times n$ grid (convex regression); (iv) $\mathcal{K}$ consists of matrices each row of which is formed by uniform sampling (with step size $1/T$) of a univariate Lipschitz function. In all these situations, we show that $A*$ can be reliably estimated for values of $T$ much smaller than what is needed for the unconstrained setting.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (44)
  1. Improved algorithms for linear stochastic bandits. In Advances in Neural Information Processing Systems, volume 24, 2011.
  2. Living on the edge: phase transitions in convex programs with random data. Information and Inference: A Journal of the IMA, 3(3):224–294, 2014.
  3. Low rank and structured modeling of high-dimensional vector autoregressions. Trans. Sig. Proc., 67(5):1207–1222, 2019.
  4. Regularized estimation in sparse high-dimensional time series models. The Annals of Statistics, 43(4):1535–1567, 2015.
  5. Pierre C. Bellec. Sharp oracle inequalities for Least Squares estimators in shape restricted regression. The Annals of Statistics, 46(2):745 – 780, 2018.
  6. Near-ideal model selection by ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT minimization. The Annals of Statistics, 37(5A):2145 – 2177, 2009.
  7. Near-optimal signal recovery from random projections: Universal encoding strategies? IEEE Transactions on Information Theory, 52(12):5406–5425, 2006.
  8. The convex geometry of linear inverse problems. Found. Comput. Math., 12(6):805–849, 2012.
  9. On risk bounds in isotonic and other shape restricted regression problems. The Annals of Statistics, 43(4):1774 – 1800, 2015.
  10. Sourav Chatterjee. A new perspective on least squares under convex constraint. The Annals of Statistics, 42(6):2340–2381, 2014.
  11. Multivariate convex regression at scale. arXiv: 2005.11588, 2021.
  12. I.C. Demetriou and P. Tzitziris. Infant mortality and economic growth: Modeling by increasing returns and least squares. In Proceedings of the World Congress on Engineering, 2017.
  13. Sjoerd Dirksen. Tail bounds via generic chaining. Electronic Journal of Probability, 20(none):1 – 29, 2015.
  14. R.M Dudley. The sizes of compact subsets of Hilbert space and continuity of Gaussian processes. Journal of Functional Analysis, 1(3):290–330, 1967.
  15. Learning sparse dynamical systems from a single sample trajectory. In 2019 IEEE 58th Conference on Decision and Control (CDC), pages 2682–2689, 2019.
  16. Y. Gordon. On Milman’s inequality and random subspaces which escape through a mesh in Rnsuperscript𝑅𝑛R^{n}italic_R start_POSTSUPERSCRIPT italic_n end_POSTSUPERSCRIPT. In Geometric Aspects of Functional Analysis, pages 84–106, 1988.
  17. Nonparametric Shape-Restricted Regression. Statistical Science, 33(4):568 – 594, 2018.
  18. Clifford Hildreth. Point estimates of ordinates of concave functions. Journal of the American Statistical Association, 49(267):598–619, 1954.
  19. Finite-time identification of stable linear systems optimality of the least-squares estimator. In 2020 59th IEEE Conference on Decision and Control (CDC), pages 996–1001, 2020.
  20. Oracle inequalities for high dimensional vector autoregressions. Journal of Econometrics, 186(2):325–344, 2015.
  21. Suprema of chaos processes and the restricted isometry property. Communications on Pure and Applied Mathematics, 67(11):1877–1904, 2014.
  22. Convex regression in multidimensions: Suboptimality of least squares estimators. arXiv: 2006.02044, 2020.
  23. High-dimensional regression with noisy and missing data: Provable guarantees with nonconvexity. The Annals of Statistics, 40(3):1637 – 1664, 2012.
  24. Linear convergence of gradient methods for estimating structured transition matrices in high-dimensional vector autoregressive models. In Advances in Neural Information Processing Systems, volume 34, pages 16751–16763, 2021.
  25. Estimating structured vector autoregressive models. In Proceedings of The 33rd International Conference on Machine Learning, pages 830–839, 2016.
  26. Shahar Mendelson. Learning without concentration. In Proceedings of The 27th Conference on Learning Theory, volume 35, pages 25–39, 2014.
  27. A unified framework for high-dimensional analysis of m-estimators with decomposable regularizers. Statistical Science, 27(4):538–557, 2012.
  28. Matey Neykov. Gaussian regression with convex constraints. In Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, volume 89, pages 31–38, 2019.
  29. Shao Qi-Man Pena Victor H., Lai Tze L. Self-Normalized Processes Limit Theory and Statistical Applications. Probability and Its Applications. Springer Berlin Heidelberg, 2009.
  30. Learning networks of stochastic differential equations. In Advances in Neural Information Processing Systems, volume 23, 2010.
  31. The generalized Lasso with non-linear observations. IEEE Trans. Inf. Theor., 62(3):1528–1537, 2016.
  32. On sparse reconstruction from Fourier and Gaussian measurements. Communications on Pure and Applied Mathematics, 61(8):1025–1045, 2008.
  33. Near optimal finite time identification of arbitrary linear dynamical systems. In Proceedings of the 36th International Conference on Machine Learning, ICML, volume 97, pages 5610–5618, 2019.
  34. Nonparametric least squares estimation of a multivariate convex regression function. The Annals of Statistics, 39(3):1633 – 1657, 2011.
  35. Finite time identification in unstable linear systems. Automatica, 96:342–353, 2018.
  36. Learning without mixing: Towards a sharp analysis of linear system identification. In Proceedings of the 31st Conference On Learning Theory, volume 75, pages 439–473, 2018.
  37. Large vector auto regressions. arXiv:1106.3915, 2011.
  38. Mihailo Stojnic. Various thresholds for ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT-optimization in compressed sensing. arXiv:0907.3666, 2009.
  39. Michel Talagrand. Upper and lower bounds for stochastic processes, volume 60. Springer, 2014.
  40. Joel A. Tropp. Convex Recovery of a Structured Signal from Independent Random Linear Measurements, pages 67–101. 2015.
  41. Roman Vershynin. Introduction to the non-asymptotic analysis of random matrices, page 210–268. Cambridge University Press, 2012.
  42. Roman Vershynin. High-Dimensional Probability: An Introduction with Applications in Data Science. Number 47 in Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, 2018.
  43. Di Wang and Ruey S. Tsay. Rate-optimal robust estimation of high-dimensional vector autoregressive models. The Annals of Statistics, 51(2):846 – 877, 2023.
  44. Finite-time analysis of vector autoregressive models under linear restrictions. Biometrika, 108(2):469–489, 2021.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets