Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Approximation by non-symmetric networks for cross-domain learning (2305.03890v3)

Published 6 May 2023 in cs.LG, cs.NA, and math.NA

Abstract: For the past 30 years or so, machine learning has stimulated a great deal of research in the study of approximation capabilities (expressive power) of a multitude of processes, such as approximation by shallow or deep neural networks, radial basis function networks, and a variety of kernel based methods. Motivated by applications such as invariant learning, transfer learning, and synthetic aperture radar imaging, we initiate in this paper a general approach to study the approximation capabilities of kernel based networks using non-symmetric kernels. While singular value decomposition is a natural instinct to study such kernels, we consider a more general approach to include the use of a family of kernels, such as generalized translation networks (which include neural networks and translation invariant kernels as special cases) and rotated zonal function kernels. Naturally, unlike traditional kernel based approximation, we cannot require the kernels to be positive definite. In particular, we obtain estimates on the accuracy of uniform approximation of functions in a Sobolev class by ReLU$r$ networks when $r$ is not necessarily an integer. Our general results apply to the approximation of functions with small smoothness compared to the dimension of the input space.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (41)
  1. F. Bach. Breaking the curse of dimensionality with convex neural networks. The Journal of Machine Learning Research, 18(1):629–681, 2017.
  2. A. R. Barron. Universal approximation bounds for superpositions of a sigmoidal function. Information Theory, IEEE Transactions on, 39(3):930–945, 1993.
  3. Understanding neural networks with reproducing kernel banach spaces. Applied and Computational Harmonic Analysis, 62:194–236, 2023.
  4. Higher transcendental functions, volume 2. McGraw-Hill New York, 1955.
  5. To understand deep learning we need to understand kernel learning. In Proceedings of the 35th International Conference on Machine Learning (PMLR), pages 80:541–549, 2018.
  6. M. Belkin and P. Niyogi. Semi-supervised learning on Riemannian manifolds. Machine learning, 56(1-3):209–239, 2004.
  7. M. Belkin and P. Niyogi. Towards a theoretical foundation for Laplacian-based manifold methods. Journal of Computer and System Sciences, 74(8):1289–1308, 2008.
  8. Concentration inequalities: A nonasymptotic theory of independence. Oxford university press, 2013.
  9. Efficient approximation of deep ReLU networks for functions on low dimensional manifolds. arXiv preprint arXiv:1908.01842, 2019.
  10. M. Cheney and B. Borden. Fundamentals of radar imaging. SIAM, 2009.
  11. Deep nets for local manifold learning. Frontiers in Applied Mathematics and Statistics, 4:12, 2018.
  12. Bigeometric organization of deep nets. Applied and Computational Harmonic Analysis, 44(3):774–785, 2018.
  13. Locally learning biomedical data using diffusion frames. Journal of Computational Biology, 19(11):1251–1264, 2012.
  14. Testing the manifold hypothesis. Journal of the American Mathematical Society, 29(4):983–1049, 2016.
  15. F. Filbir and H. N. Mhaskar. A quadrature formula for diffusion polynomials corresponding to a generalized heat kernel. Journal of Fourier Analysis and Applications, 16(5):629–657, 2010.
  16. F. Filbir and H. N. Mhaskar. Marcinkiewicz–Zygmund measures on manifolds. Journal of Complexity, 27(6):568–596, 2011.
  17. J. Friedman and J.-P. Tillich. Wave equations for graphs and the edge-based laplacian. Pacific Journal of Mathematics, 216(2):229–266, 2004.
  18. Quadrature in Besov spaces on the euclidean sphere. Journal of Complexity, 23(4):528–552, 2007.
  19. Uniform approximation by neural networks activated by first and second order ridge splines. arXiv preprint arXiv:1607.07819, 2016.
  20. V. Kůrková and M. Sanguineti. Bounds on rates of variable basis and neural network approximation. IEEE Transactions on Information Theory, 47(6):2659–2665, 2001.
  21. V. Kůrková and M. Sanguineti. Comparison of worst case errors in linear and neural network approximation. IEEE Transactions on Information Theory, 48(1):264–275, 2002.
  22. S. S. Lafon. Diffusion maps and geometric harmonics. PhD thesis, Yale University, Yale, 2004.
  23. M. Maggioni and H. N. Mhaskar. Diffusion polynomial frames on metric measure spaces. Applied and Computational Harmonic Analysis, 24(3):329–353, 2008.
  24. H. Mhaskar and J. Prestin. Polynomial frames for the detection of singularities. Wavelet analysis and multiresolution methods, 212:273–298, 2000.
  25. H. N. Mhaskar. Weighted quadrature formulas and approximation by zonal function networks on the sphere. Journal of Complexity, 22(3):348–370, 2006.
  26. H. N. Mhaskar. Eignets for function approximation on manifolds. Applied and Computational Harmonic Analysis, 29(1):63–87, 2010.
  27. H. N. Mhaskar. A generalized diffusion frame for parsimonious representation of functions on data defined manifolds. Neural Networks, 24(4):345–359, 2011.
  28. H. N. Mhaskar. A unified framework for harmonic analysis of functions on directed graphs and changing data. Appl. Comput. Harm. Anal., 44(3):611–644, 2018.
  29. H. N. Mhaskar. Dimension independent bounds for general shallow networks. Neural Networks, 123:142–152, 2020.
  30. H. N. Mhaskar. Kernel-based analysis of massive data. Frontiers in Applied Mathematics and Statistics, 6:30, 2020.
  31. H. N. Mhaskar. Function approximation with zonal function networks with activation functions analogous to the rectified linear unit functions. Journal of Complexity, 51:1–19, April 2019.
  32. A witness function based construction of discriminative models using hermite polynomials. Frontiers in Applied Mathematics and Statistics, 6:31, 2020.
  33. Degree of approximation by neural and translation networks with a single hidden layer. Advances in Applied Mathematics, 16(2):151–183, 1995.
  34. Approximation properties of zonal function networks using scattered data on the sphere. Advances in Computational Mathematics, 11(2-3):121–137, 1999.
  35. H. N. Mhaskar and T. Poggio. Deep vs. shallow networks: An approximation theory perspective. Analysis and Applications, 14(06):829–848, 2016.
  36. C. Müller. Spherical harmonics, volume 17. Springer, 2006.
  37. T. Poggio and F. Anselmi. Visual cortex and deep networks: learning invariant representations. MIT Press, 2016.
  38. A. Rahimi and B. Recht. Random features for large-scale kernel machines. Advances in neural information processing systems, 20, 2007.
  39. J. Schmidt-Hieber. Deep ReLU network approximation of functions on a manifold. arXiv preprint arXiv:1908.00695, 2019.
  40. A. Singer. From graph to manifold Laplacian: The convergence rate. Applied and Computational Harmonic Analysis, 21(1):128–134, 2006.
  41. N. I. Vilenkin. Special functions and the theory of group representations, volume 22. American Mathematical Soc., 1978.
Citations (1)

Summary

We haven't generated a summary for this paper yet.