Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sparse Representer Theorems for Learning in Reproducing Kernel Banach Spaces (2305.12584v2)

Published 21 May 2023 in math.FA and cs.LG

Abstract: Sparsity of a learning solution is a desirable feature in machine learning. Certain reproducing kernel Banach spaces (RKBSs) are appropriate hypothesis spaces for sparse learning methods. The goal of this paper is to understand what kind of RKBSs can promote sparsity for learning solutions. We consider two typical learning models in an RKBS: the minimum norm interpolation (MNI) problem and the regularization problem. We first establish an explicit representer theorem for solutions of these problems, which represents the extreme points of the solution set by a linear combination of the extreme points of the subdifferential set, of the norm function, which is data-dependent. We then propose sufficient conditions on the RKBS that can transform the explicit representation of the solutions to a sparse kernel representation having fewer terms than the number of the observed data. Under the proposed sufficient conditions, we investigate the role of the regularization parameter on sparsity of the regularized solutions. We further show that two specific RKBSs: the sequence space $\ell_1(\mathbb{N})$ and the measure space can have sparse representer theorems for both MNI and regularization models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. When is there a representer theorem? Vector versus matrix regularizers. Journal of Machine Learning Research, 10:2507–2529, 2009.
  2. F. Bach. Breaking the curse of dimensionality with convex neural networks. Journal of Machine Learning Research, 18:629–681, 2017.
  3. Understanding neural networks with reproducing kernel Banach spaces. Applied and Computational Harmonic Analysis, 62:194–236, 2023.
  4. On representer theorems and convex regularization. SIAM Journal on Optimization, 29:1260–1281, 2019.
  5. K. Bredies and M. Carioni. Sparsity of solutions for variational inverse problems with finite-dimensional data. Calculus of Variations and Partial Differential Equations, 59:1–26, 2020.
  6. A duality approach to regularized learning problems in Banach spaces. Journal of Complexity, 81:101818, 2024.
  7. R. Cheng and Y. Xu. Minimum norm interpolation in the ℓ1⁢(ℕ)subscriptℓ1ℕ\ell_{1}(\mathbb{N})roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ( blackboard_N ) space. Analysis and Applications, 19:21–42, 2021.
  8. I. Cioranescu. Geometry of Banach Spaces, Duality Mappings and Nonlinear Problems. Mathematics and its Applications, volume 62. Kluwer Academic Publishers Group, Dordrecht, 1990.
  9. J. B. Conway. A Course in Functional Analysis. 2nd Edition. Springer-Verlag, New York, 1990.
  10. D. D. Cox and F. O’Sullivan. Asymptotic analysis of penalized likelihood and related estimators. The Annals of Statistics, 18:1676–1695, 1990.
  11. C. deBoor and R. E. Lynch. On splines and their minimum properties. Journal of Mathematics and Mechanics, 15:953–969, 1966.
  12. L. E. Dubins. On extreme points of convex sets. Journal of Mathematical Analysis and Applications, 5:237–244, 1962.
  13. Solving support vector machines in reproducing kernel Banach spaces with positive definite functions. Applied and Computational Harmonic Analysis, 38:115–139, 2015.
  14. The future of deep learning will be sparse. SIAM News, May, 3, 2021.
  15. Generalized representer theorems in Banach spaces. Analysis and Applications, 19:125–146, 2021.
  16. G. Kimeldorf and G. Wahba. A correspondence between bayesian estimation on stochastic processes and smoothing by splines. Annals of Mathematical Statistics, 41:495–502, 1970.
  17. A. J. Kurdila and M. Zabarankin. Convex Functional Analysis. Systems & Control: Foundations & Applications. Birkhäuser Verlag, Basel, 2005.
  18. Multi-task learning in vector-valued reproducing kernel Banach spaces with the ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT norm. Journal of Complexity, 63:101514, 2021.
  19. R. Lin and Y. Xu. Reproducing kernel Banach spaces of bounded continuous functions. Preparation, 2022.
  20. On reproducing kernel Banach spaces: generic definitions and unified framework of constructions. Acta Mathematica Sinica, English Series, 38:1459–1483, 2022.
  21. Sparse regularized learning in the reproducing kernel Banach spaces with the ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT norm. Mathematical Foundations of Computing, 3:205–218, 2020.
  22. Parameter choices for sparse regularization with the ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT norm. Inverse Problems, 39:025004, 2023.
  23. R. E. Megginson. An Introduction to Banach Space Theory. Graduate Texts in Mathematics, volume 183. Springer-Verlag, New York, 1998.
  24. C. A. Micchelli and M. Pontil. A function representation for learning in Banach spaces. In International Conference on Computational Learning Theory, pages 255–269. Springer, 2004.
  25. R. Parhi and R. D. Nowak. Banach space representer theorems for neural networks and ridge splines. Journal of Machine Learning Research, 22:1960–1999, 2021.
  26. ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT regularization in infinite dimensional feature spaces. In International Conference on Computational Learning Theory, pages 544–558. Springer, 2007.
  27. A generalized representer theorem. In Proceeding of the Fourteenth Annual Conference on Computational Learning Theory and the Fifth European Conference on Computational Learning Theory, pages 416–426. Springer-Verlag, London, 2001.
  28. Vector-valued variation spaces and width bounds for DNNs: insights on weight decay regularization. arXiv preprint arXiv:2305.16534, 2023.
  29. Reproducing kernel Banach spaces with the ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT norm. Applied and Computational Harmonic Analysis, 34:96–116, 2013.
  30. Duality for neural networks through reproducing kernel Banach spaces. arXiv preprint arXiv:2211.05020, 2023.
  31. M. Unser. A unifying representer theorem for inverse problems and machine learning. Foundations of Computational Mathematics, 21:941–960, 2021.
  32. M. Unser and S. Aziznejad. Convex optimization in sums of Banach spaces. Applied and Computational Harmonic Analysis, 56:1–25, 2022.
  33. Representer theorems for sparsity-promoting ℓ1subscriptℓ1\ell_{1}roman_ℓ start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT regularization. IEEE Transactions on Information Theory, 62:5167–5180, 2016.
  34. Splines are universal solutions of linear inverse problems with generalized TV regularization. SIAM Review, 59:769–793, 2017.
  35. R. Wang and Y. Xu. Representer theorems in Banach spaces: minimum norm interpolation, regularized learning and semi-discrete inverse problems. Journal of Machine Learning Research, 22:1–65, 2021.
  36. Y. Xu. Sparse machine learning in Banach spaces. Applied Numerical Mathematics, 187:138–157, 2023.
  37. Y. Xu and Q. Ye. Generalized Mercer kernels and reproducing kernel Banach spaces. Memoirs of the American Mathematical Society, 258(1243), 2019.
  38. Reproducing kernel Banach spaces for machine learning. Journal of Machine Learning Research, 10:2741–2775, 2009.
  39. H. Zhang and J. Zhang. Regularized learning in Banach spaces as an optimization problem: representer theorems. Journal of Global Optimization, 54:235–250, 2012.
  40. C. Zălinescu. Convex Analysis in General Vector Spaces. World Scientific, New Jersey, 2002.
Citations (5)

Summary

We haven't generated a summary for this paper yet.