Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient online learning with kernels for adversarial large scale problems (1902.09917v2)

Published 26 Feb 2019 in stat.ML, cs.LG, math.ST, and stat.TH

Abstract: We are interested in a framework of online learning with kernels for low-dimensional but large-scale and potentially adversarial datasets. We study the computational and theoretical performance of online variations of kernel Ridge regression. Despite its simplicity, the algorithm we study is the first to achieve the optimal regret for a wide range of kernels with a per-round complexity of order $n\alpha$ with $\alpha < 2$. The algorithm we consider is based on approximating the kernel with the linear span of basis functions. Our contributions is two-fold: 1) For the Gaussian kernel, we propose to build the basis beforehand (independently of the data) through Taylor expansion. For $d$-dimensional inputs, we provide a (close to) optimal regret of order $O((\log n){d+1})$ with per-round time complexity and space complexity $O((\log n){2d})$. This makes the algorithm a suitable choice as soon as $n \gg ed$ which is likely to happen in a scenario with small dimensional and large-scale dataset; 2) For general kernels with low effective dimension, the basis functions are updated sequentially in a data-adaptive fashion by sampling Nystr{\"o}m points. In this case, our algorithm improves the computational trade-off known for online kernel regression.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Rémi Jézéquel (4 papers)
  2. Pierre Gaillard (44 papers)
  3. Alessandro Rudi (70 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.