Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Semi-supervised Fréchet Regression (2404.10444v1)

Published 16 Apr 2024 in math.ST, cs.LG, stat.ML, and stat.TH

Abstract: This paper explores the field of semi-supervised Fr\'echet regression, driven by the significant costs associated with obtaining non-Euclidean labels. Methodologically, we propose two novel methods: semi-supervised NW Fr\'echet regression and semi-supervised kNN Fr\'echet regression, both based on graph distance acquired from all feature instances. These methods extend the scope of existing semi-supervised Euclidean regression methods. We establish their convergence rates with limited labeled data and large amounts of unlabeled data, taking into account the low-dimensional manifold structure of the feature space. Through comprehensive simulations across diverse settings and applications to real data, we demonstrate the superior performance of our methods over their supervised counterparts. This study addresses existing research gaps and paves the way for further exploration and advancements in the field of semi-supervised Fr\'echet regression.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. A framework for learning predictive structures from multiple tasks and unlabeled data. Journal of Machine Learning Research, 6(11).
  2. Unconstrained and curvature-constrained shortest-path distances and their approximation. Discrete & Computational Geometry, 62:1–28.
  3. Density-sensitive semisupervised inference. The Annals of Statistics, 41(2):751 – 771.
  4. Semi-supervised linear regression. Journal of the American Statistical Association, 117(540):2238–2251.
  5. Semi-supervised learning on riemannian manifolds. Machine learning, 56:209–239.
  6. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of machine learning research, 7(11).
  7. Local polynomial regression on unknown manifolds. Lecture Notes-Monograph Series, pages 177–186.
  8. Geometry of the space of phylogenetic trees. Advances in Applied Mathematics, 27(4):733–767.
  9. Combining labeled and unlabeled data with co-training. In Proceedings of the eleventh annual conference on Computational learning theory, pages 92–100.
  10. Deep neural networks for learning graph representations. In Proceedings of the AAAI conference on artificial intelligence, volume 30.
  11. Efficient and adaptive linear regression in semi-supervised settings. The Annals of Statistics, 46(4):1541 – 1572.
  12. Wasserstein regression. Journal of the American Statistical Association, 118(542):869–882.
  13. Uniform convergence of local fréchet regression with applications to locating extrema and time warping for metric space valued trajectories. The Annals of Statistics, 50(3):1573–1592.
  14. Regression models on riemannian symmetric spaces. Journal of the Royal Statistical Society Series B: Statistical Methodology, 79(2):463–482.
  15. Choosing the most relevant level sets for depicting a sample of densities. Computational Statistics, 32:1083–1113.
  16. Nonparametric regression for spherical data. Journal of the American Statistical Association, 109(506):748–763.
  17. Intrinsic persistent homology via density-based metric learning. Journal of Machine Learning Research, 24(75):1–42.
  18. Fréchet, M. (1948). Les éléments aléatoires de nature quelconque dans un espace distancié. In Annales de l’institut Henri Poincaré, volume 10, pages 215–310.
  19. When can unlabeled data improve the learning rate? In Conference on Learning Theory, pages 1500–1518. PMLR.
  20. Shortest path through random points. The Annals of Applied Probability, 26(5):2791–2823.
  21. A simple algorithm for semi-supervised learning with improved generalization error bound.
  22. Extrinsic local regression on manifold-valued data. Journal of the American Statistical Association, 112(519):1261–1273.
  23. Lin, Z. (2019). Riemannian geometry of symmetric positive definite matrices via cholesky decomposition. SIAM Journal on Matrix Analysis and Applications, 40(4):1353–1370.
  24. Additive models for symmetric positive-definite matrices and lie groups. Biometrika, 110(2):361–379.
  25. Improved estimators for semi-supervised high-dimensional regression model. Electronic Journal of Statistics, 16(2):5437–5487.
  26. Minimax-optimal semi-supervised regression on unknown manifolds. In Artificial Intelligence and Statistics, pages 933–942. PMLR.
  27. Niyogi, P. (2013). Manifold regularization and semi-supervised learning: Some theoretical analyses. Journal of Machine Learning Research, 14(5).
  28. Fréchet regression for random objects with euclidean predictors. The Annals of Statistics, 47(2):691–719.
  29. Semi-supervised active linear regression. Advances in Neural Information Processing Systems, 35:1294–1306.
  30. Rigollet, P. (2007). Generalization error bounds in semi-supervised classification under the cluster assumption. Journal of Machine Learning Research, 8(7).
  31. Nonlinear dimensionality reduction by locally linear embedding. science, 290(5500):2323–2326.
  32. The graph neural network model. IEEE transactions on neural networks, 20(1):61–80.
  33. Schötz, C. (2022). Nonparametric regression in nonstandard spaces. Electronic Journal of Statistics, 16(2):4679–4741.
  34. Unlabeled data: Now it helps, now it doesn’t. Advances in neural information processing systems, 21.
  35. Semi-supervised learning using sparse eigenfunction bases. Advances in Neural Information Processing Systems, 22.
  36. A general m-estimation theory in semi-supervised framework. Journal of the American Statistical Association, pages 1–11.
  37. Stone, C. J. (1982). Optimal global rates of convergence for nonparametric regression. The annals of statistics, pages 1040–1053.
  38. Talagrand, M. (2014). Upper and lower bounds for stochastic processes.
  39. A global geometric framework for nonlinear dimensionality reduction. science, 290(5500):2319–2323.
  40. Large margin semi-supervised learning. Journal of Machine Learning Research, 8(8).
  41. On efficient large margin semisupervised learning: Method and theory. Journal of Machine Learning Research, 10(3).
  42. Statistical analysis of semi-supervised regression. Advances in Neural Information Processing Systems, 20.
  43. Local polynomial regression for symmetric positive definite matrices. Journal of the Royal Statistical Society Series B: Statistical Methodology, 74(4):697–719.
  44. Dimension reduction and data visualization for fréchet regression. arXiv preprint arXiv:2110.00467.
  45. Intrinsic regression models for positive-definite matrices with applications to diffusion tensor imaging. Journal of the American Statistical Association, 104(487):1203–1212.
  46. Semi-supervised learning using gaussian fields and harmonic functions. In Proceedings of the 20th International conference on Machine learning (ICML-03), pages 912–919.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com