Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Stability of convergence rates: Kernel interpolation on non-Lipschitz domains (2203.12532v2)

Published 23 Mar 2022 in math.NA and cs.NA

Abstract: Error estimates for kernel interpolation in Reproducing Kernel Hilbert Spaces (RKHS) usually assume quite restrictive properties on the shape of the domain, especially in the case of infinitely smooth kernels like the popular Gaussian kernel. In this paper we leverage an analysis of greedy kernel algorithms to prove that it is possible to obtain convergence results (in the number of interpolation points) for kernel interpolation for arbitrary domains $\Omega \subset \mathbb{R}d$, thus allowing for non-Lipschitz domains including e.g. cusps and irregular boundaries. Especially we show that, when going to a smaller domain $\tilde{\Omega} \subset \Omega \subset \mathbb{R}d$, the convergence rate does not deteriorate - i.e. the convergence rates are stable with respect to going to a subset. The impact of this result is explained on the examples of kernels of finite as well as infinite smoothness like the Gaussian kernel. A comparison to approximation in Sobolev spaces is drawn, where the shape of the domain $\Omega$ has an impact on the approximation properties. Numerical experiments illustrate and confirm the experiments.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Tizian Wenzel (17 papers)
  2. Gabriele Santin (42 papers)
  3. Bernard Haasdonk (45 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.