Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
166 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning on manifolds without manifold learning (2402.12687v2)

Published 20 Feb 2024 in cs.LG and stat.ML

Abstract: Function approximation based on data drawn randomly from an unknown distribution is an important problem in machine learning. The manifold hypothesis assumes that the data is sampled from an unknown submanifold of a high dimensional Euclidean space. A great deal of research deals with obtaining information about this manifold, such as the eigendecomposition of the Laplace-Beltrami operator or coordinate charts, and using this information for function approximation. This two-step approach implies some extra errors in the approximation stemming from estimating the basic quantities of the data manifold in addition to the errors inherent in function approximation. In this paper, we project the unknown manifold as a submanifold of an ambient hypersphere and study the question of constructing a one-shot approximation using a specially designed sequence of localized spherical polynomial kernels on the hypersphere. Our approach does not require preprocessing of the data to obtain information about the manifold other than its dimension. We give optimal rates of approximation for relatively ``rough'' functions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Hrushikesh N Mhaskar “Dimension independent bounds for general shallow networks” In Neural Networks 123 Elsevier, 2020, pp. 142–152
  2. Hrushikesh N Mhaskar “Function approximation with zonal function networks with activation functions analogous to the rectified linear unit functions” In Journal of Complexity 51, April 2019, pp. 1–19
  3. “Laplacian eigenmaps for dimensionality reduction and data representation” In Neural computation 15.6 MIT Press, 2003, pp. 1373–1396
  4. “Towards a theoretical foundation for Laplacian-based manifold methods” In Journal of Computer and System Sciences 74.8 Elsevier, 2008, pp. 1289–1308
  5. A. Singer “From graph to manifold Laplacian: The convergence rate” Special Issue: Diffusion Maps and Wavelets In Applied and Computational Harmonic Analysis 21.1, 2006, pp. 128–134
  6. “Special issue: Diffusion maps and wavelets” In Appl. and Comput. Harm. Anal. 21.1, 2006
  7. “Bigeometric organization of deep nets” In Applied and Computational Harmonic Analysis 44.3 Elsevier, 2018, pp. 774–785
  8. Johannes Schmidt-Hieber “Deep ReLU network approximation of functions on a manifold” In arXiv preprint arXiv:1908.00695, 2019
  9. Barak Sober, Yariv Aizenbud and David Levin “Approximation of functions over manifolds: A Moving Least-Squares approach” In Journal of Computational and Applied Mathematics 383, 2021, pp. 113140
  10. “Local Linear Regression on Manifolds and Its Geometric Interpretation” In Journal of the American Statistical Association 108.504 [American Statistical Association, Taylor & Francis, Ltd.], 2013, pp. 1421–1434
  11. Charles K. Chui and Hrushikesh N. Mhaskar “Deep Nets for Local Manifold Learning” In Frontiers in Applied Mathematics and Statistics 4, 2018, pp. 12 DOI: 10.3389/fams.2018.00012
  12. “Diffusion polynomial frames on metric measure spaces” In Applied and Computational Harmonic Analysis 24.3 Elsevier, 2008, pp. 329–353
  13. M. Ehler, F. Filbir and H.N. Mhaskar “Locally Learning Biomedical Data Using Diffusion Frames” In Journal of Computational Biology 19.11 Mary Ann Liebert, Inc. 140 Huguenot Street, 3rd Floor New Rochelle, NY 10801 USA, 2012, pp. 1251–1264
  14. “Marcinkiewicz–Zygmund measures on manifolds” In Journal of Complexity 27.6 Elsevier, 2011, pp. 568–596
  15. Hrushikesh Mhaskar “Kernel-Based Analysis of Massive Data” In Frontiers in Applied Mathematics and Statistics 6, 2020, pp. 30
  16. Hrushikesh Narhar Mhaskar “A generalized diffusion frame for parsimonious representation of functions on data defined manifolds” In Neural Networks 24.4 Elsevier, 2011, pp. 345–359
  17. Hrushikesh Narhar Mhaskar “Eignets for function approximation on manifolds” In Applied and Computational Harmonic Analysis 29.1 Elsevier, 2010, pp. 63–87
  18. F. Girosi, M.B. Jones and T. Poggio “Regularization theory and neural networks architectures” In Neural computation 7.2 MIT Press, 1995, pp. 219–269
  19. H.N. Mhaskar “A direct approach for function approximation on data defined manifolds” In Neural Networks 132, 2020, pp. 253–268
  20. H.N. Mhaskar, S.V. Pereverzyev and M.D. Walt “A Function Approximation Approach to the Prediction of Blood Glucose Levels” In Frontiers in Applied Mathematics and Statistics 7, 2021, pp. 53 DOI: 10.3389/fams.2021.707884
  21. Eric Mason, Hrushikesh Narhar Mhaskar and Adam Guo “A manifold learning approach for gesture identification from micro-Doppler radar measurements” arXiv preprint arXiv:2110.01670, 2021 In Neural Networks 152, 2022, pp. 353–369
  22. W. Gautschi “Orthogonal polynomials: computation and approximation” Oxford University Press on Demand, 2004
  23. Claus Müller “Spherical harmonics” Springer, 2006
  24. Elias M Stein and Guido Weiss “Introduction to Fourier analysis on Euclidean spaces (PMS-32)” Princeton university press, 2016
  25. G. Szegö “Orthogonal Polynomials”, American Math. Soc: Colloquium publ American Mathematical Society, 1975
  26. Richard Askey “Orthogonal Polynomials and Special Functions” Society for IndustrialApplied Mathematics, 1975 DOI: 10.1137/1.9781611970470
  27. H.N. Mhaskar “Polynomial operators and local smoothness classes on the unit interval” In Journal of Approximation Theory 131.2, 2004, pp. 243–267
  28. H.N. Mhaskar “On the representation of smooth functions on the sphere using finitely many bits” In Applied and Computational Harmonic Analysis 18.3, 2005, pp. 215–233
  29. Kh P Rustamov “ON APPROXIMATION OF FUNCTIONS ON THE SPHERE” In Izvestiya: Mathematics 43.2, 1994, pp. 311
  30. “Approximation Theory and Harmonic Analysis on Spheres and Balls”, Springer Monographs in Mathematics Springer New York, 2013
  31. “Input layer regularization for magnetic resonance relaxometry biexponential parameter estimation” In Magnetic Resonance in Chemistry 60(11), 2022, pp. 1076–1086
  32. Stéphane Boucheron, Gábor Lugosi and Pascal Massart “Concentration Inequalities” Oxford University Press, 2013
  33. Manfredo P. Carmo “Riemannian Geometry” Birkhäuser, 1992
  34. W.M. Boothby “An Introduction to Differentiable Manifolds and Riemannian Geometry: An Introduction to Differentiable Manifolds and Riemannian Geometry”, ISSN Elsevier Science, 1975
  35. “Differential Topology” AMS Chelsea Pub., 2010
  36. John Roe “Elliptic operators, topology and asymptotic methods” Addison Wesley Longman Inc., 1998
  37. “Fundamentals of Approximation Theory” New Dehli, India: Narosa Publishing House, 2000
  38. “Semi-Supervised Learning on Riemannian Manifolds” In Machine Learning 56, 2004, pp. 209–239

Summary

  • The paper introduces a novel method for direct function approximation on manifolds using spherical polynomials, eliminating the need for manifold learning.
  • It establishes optimal error bounds that depend on the intrinsic dimensionality, effectively mitigating the curse of dimensionality.
  • The approach enhances computational efficiency and robustness in applications such as computer vision and signal processing by leveraging low-dimensional data structures.

An Overview of "Learning on Manifolds without Manifold Learning"

The paper "Learning on Manifolds without Manifold Learning" by H. N. Mhaskar and Ryan O'Dowd addresses a pivotal challenge in the field of machine learning—function approximation on high-dimensional data sampled from a low-dimensional manifold. Traditional methods rely on manifold learning to derive characteristics of the manifold before performing function approximation, typically as a two-step process. These traditional approaches entail manifold learning efforts such as eigendecomposition of the Laplace-Beltrami operator or deriving coordinate charts, imposing an extra error due to manifold approximation.

Overview of the Methodology

The authors propose a method that avoids the manifold learning step altogether by directly targeting function approximation using spherical polynomials. This approach circumvents the need to explicitly learn manifold properties other than its dimension. The data is instead viewed as samples drawn from a sub-manifold of an ambient hypersphere, which allows the use of spherical polynomials for constructing a one-shot approximation. This methodology is rooted in approximation theory and leverages optimal error bounds assuming the manifold structure of the data.

Specifically, the paper introduces a kernel-based approximation on hyperspheres with a direct sum of spherical harmonics up to a certain degree. The polynomial expansion's coefficients are computed directly from the data, without the need to infer the manifold’s underlying structure.

Theoretical Framework and Results

The primary contribution of the paper is a theoretical framework that establishes optimal rates of approximation for relatively rough functions using spherical harmonics. The results show that for a function f(x)=E(zy)f(x) = \mathbb{E}(z|y), where (yj,zj)(y_j, z_j) are samples, the approximation error Fn(D;)ff0X\| F_n(\mathcal{D}; \cdot) - ff_0 \|_{\mathbb{X}} scales as:

(f0X+zX×Ω+(ff0)Wγ(X))(log(M/δ)q+2γM)γ/(q+2γ)\lesssim \left( \sqrt{f_0}_{\mathbb{X}} + z_{\mathbb{X} \times \Omega} + (ff_0)_{W_\gamma(\mathbb{X})} \right) \left( \frac{\log(M/\delta)^{q+2\gamma}}{M} \right)^{\gamma/(q+2\gamma)}

Where f0f_0 represents the density function of the manifold distribution, and MM is the sample size. The approach demonstrates that the rate of convergence depends on the dimensionality of the manifold rather than the ambient space—effectively mitigating the curse of dimensionality.

Implications and Future Directions

The implications of this research are significant in machine learning and artificial intelligence, particularly in applications where data do not densely populate the ambient space but adhere to lower-dimensional manifolds, such as computer vision and signal processing. By eliminating the manifold learning step, the method becomes computationally more efficient and potentially more robust to diverse data distributions.

The notable outcomes invite several avenues for future research, including:

  • Extending approximation techniques to cover more complex or non-compact manifolds.
  • Exploring the integration of this approach within neural network architectures to enhance feature learning on manifold-structured data.
  • Investigating the interplay of this direct approximation method with probabilistic modeling and inference to handle uncertainties in data distributions.

Overall, this paper presents a methodologically sound and theoretically supported alternative for manifold-related learning tasks, highlighting its potential to streamline and improve the efficiency of machine learning workflows that involve high-dimensional data approximation on unknown low-dimensional sub-manifolds.