Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Optimal Rates for Spectral Algorithms with Least-Squares Regression over Hilbert Spaces (1801.06720v4)

Published 20 Jan 2018 in stat.ML, cs.LG, and math.FA

Abstract: In this paper, we study regression problems over a separable Hilbert space with the square loss, covering non-parametric regression over a reproducing kernel Hilbert space. We investigate a class of spectral/regularized algorithms, including ridge regression, principal component regression, and gradient methods. We prove optimal, high-probability convergence results in terms of variants of norms for the studied algorithms, considering a capacity assumption on the hypothesis space and a general source condition on the target function. Consequently, we obtain almost sure convergence results with optimal rates. Our results improve and generalize previous results, filling a theoretical gap for the non-attainable cases.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Junhong Lin (29 papers)
  2. Alessandro Rudi (70 papers)
  3. Lorenzo Rosasco (144 papers)
  4. Volkan Cevher (216 papers)
Citations (89)

Summary

We haven't generated a summary for this paper yet.