Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
120 tokens/sec
GPT-4o
10 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
3 tokens/sec
DeepSeek R1 via Azure Pro
51 tokens/sec
2000 character limit reached

Koopman Theory and Linear Approximation Spaces (1811.10809v1)

Published 27 Nov 2018 in math.DS and math.NA

Abstract: Koopman theory studies dynamical systems in terms of operator theoretic properties of the Perron-Frobenius operator $\mathcal{P}$ and Koopman operator $\mathcal{U}$ respectively. In this paper, we derive the rates of convergence of approximations of $\mathcal{P}$ or $\mathcal{U}$ that are generated by finite dimensional bases like wavelets, multiwavelets, and eigenfunctions, as well as approaches that use samples of the input and output of the system in conjunction with these bases. We introduce a general class of priors that describe the information available for constructing such approximations and facilitate the error estimates in many applications of interest. These priors are defined in terms of the action of $\mathcal{P}$ or $\mathcal{U}$ on certain linear approximation spaces. The rates of convergence for the estimates of these operators are investigated under a variety of situations that are motivated from associated assumptions in practical applications. When the estimates of these operators are generated by samples, it is shown that the error in approximation of the Perron-Frobenius or Koopman operators can be decomposed into two parts, the approximation error and the sampling error. This result emphasizes that sample-based estimates of Perron-Frobenius and Koopman operators are subject to the well-known trade-off between the bias and variance that contribute to the error, a balance that also features in nonlinear regression and statistical learning theory.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.