Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficiently Learning the Graph for Semi-supervised Learning (2306.07098v1)

Published 12 Jun 2023 in cs.LG and cs.AI

Abstract: Computational efficiency is a major bottleneck in using classic graph-based approaches for semi-supervised learning on datasets with a large number of unlabeled examples. Known techniques to improve efficiency typically involve an approximation of the graph regularization objective, but suffer two major drawbacks - first the graph is assumed to be known or constructed with heuristic hyperparameter values, second they do not provide a principled approximation guarantee for learning over the full unlabeled dataset. Building on recent work on learning graphs for semi-supervised learning from multiple datasets for problems from the same domain, and leveraging techniques for fast approximations for solving linear systems in the graph Laplacian matrix, we propose algorithms that overcome both the above limitations. We show a formal separation in the learning-theoretic complexity of sparse and dense graph families. We further show how to approximately learn the best graphs from the sparse families efficiently using the conjugate gradient method. Our approach can also be used to learn the graph efficiently online with sub-linear regret, under mild smoothness assumptions. Our online learning results are stated generally, and may be useful for approximate and efficient parameter tuning in other problems. We implement our approach and demonstrate significant ($\sim$10-100x) speedups over prior work on semi-supervised learning with learned graphs on benchmark datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Nonstochastic multi-armed bandits with graph-structured feedback. SIAM Journal on Computing, 46(6):1785–1826, 2017.
  2. Neural network learning: theoretical foundations, volume 9. Cambridge University Press, Cambridge, 1999.
  3. O. Axelsson. A class of iterative methods for finite element equations. Computer Methods in Applied Mechanics and Engineering, 9(2):123–137, 1976. ISSN 0045-7825. https://doi.org/10.1016/0045-7825(76)90056-6. URL https://www.sciencedirect.com/science/article/pii/0045782576900566.
  4. Maria-Florina Balcan. Book chapter Data-Driven Algorithm Design. In Beyond Worst Case Analysis of Algorithms, T. Roughgarden (Ed). Cambridge University Press, 2020.
  5. Learning revenue maximizing menus of lotteries and two-part tariffs. arXiv preprint arXiv:2302.11700, 2023.
  6. A discriminative model for semi-supervised learning. Journal of the ACM (JACM), 57(3):1–46, 2010.
  7. Data driven semi-supervised learning. Advances in Neural Information Processing Systems (NeurIPS), 34, 2021.
  8. Learning-theoretic foundations of algorithm configuration for combinatorial partitioning problems. In Conference on Learning Theory (COLT), pages 213–274. PMLR, 2017.
  9. Dispersion for data-driven algorithm design, online learning, and private optimization. In 2018 IEEE 59th Annual Symposium on Foundations of Computer Science (FOCS), pages 603–614. IEEE, 2018.
  10. Semi-bandit optimization in the dispersed setting. In Conference on Uncertainty in Artificial Intelligence (UAI), pages 909–918. PMLR, 2020a.
  11. Learning piecewise Lipschitz functions in changing environments. In International Conference on Artificial Intelligence and Statistics (AISTATS), pages 3567–3577. PMLR, 2020b.
  12. Refined bounds for algorithm configuration: The knife-edge of dual class approximability. In International Conference on Machine Learning (ICML), pages 580–590. PMLR, 2020c.
  13. Provably tuning the elasticnet across instances. Advances in Neural Information Processing Systems (NeurIPS), 35:27769–27782, 2022a.
  14. Faster algorithms for learning to link, align sequences, and price two-part tariffs. arXiv preprint arXiv:2204.03569, 2022b.
  15. An analysis of robustness of non-lipschitz networks. Journal of Machine Learning Research (JMLR), 2023.
  16. Generalization bounds for data-driven numerical linear algebra. In Conference on Learning Theory (COLT), pages 2013–2040. PMLR, 2022.
  17. Learning from labeled and unlabeled data using graph mincuts. In International Conference on Machine Learning (ICML), 2001.
  18. Alexandre d’Aspremont. Smooth optimization with approximate gradient. SIAM Journal on Optimization, 19(3):1171–1183, 2008.
  19. Efficient non-parametric function induction in semi-supervised learning. In International Workshop on Artificial Intelligence and Statistics (AISTATS), pages 96–103. PMLR, 2005.
  20. A PAC approach to application-specific algorithm selection. SIAM Journal on Computing, 46(3):992–1017, 2017.
  21. A survey of nonlinear conjugate gradient methods. Pacific journal of Optimization, 2(1):35–58, 2006.
  22. Methods of conjugate gradients for solving. Journal of research of the National Bureau of Standards, 49(6):409, 1952.
  23. J.J. Hull. A database for handwritten text recognition research. IEEE Transactions on Pattern Analysis and Machine Intelligence, 16(5):550–554, 1994. 10.1109/34.291440.
  24. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations (ICLR), 2017.
  25. Kdd workshop on machine learning in finance. In Conference on Knowledge Discovery and Data Mining (SIGKDD), pages 4882–4883, 2022.
  26. Industrial applications of machine learning. CRC press, 2018.
  27. Gradient-based learning applied to document recognition. IEEE, 86(11):2278–2324, 1998. 10.1109/5.726791.
  28. Large graph construction for scalable semi-supervised learning. In International conference on machine learning (ICML), 2010.
  29. Ulla Miekkala. Graph properties for splitting with grounded laplacian matrices. BIT Numerical Mathematics, 33(3):485–495, 1993.
  30. Yu E Nesterov. A method for solving the convex programming problem with convergence rate o⁢(1k2)𝑜1superscript𝑘2o(\frac{1}{k^{2}})italic_o ( divide start_ARG 1 end_ARG start_ARG italic_k start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT end_ARG ). In Dokl. Akad. Nauk SSSR,, volume 269, pages 543–547, 1983.
  31. Using the mutual k-nearest neighbor graphs for semi-supervised classification on natural language data. In Fifteenth conference on computational natural language learning (CoNLL), pages 154–162, 2011.
  32. David Pollard. Convergence of stochastic processes. Springer Science & Business Media, 2012.
  33. Towards generalization and simplicity in continuous control. Advances in Neural Information Processing Systems (NeurIPS), 30, 2017.
  34. Trust region policy optimization. In International Conference on Machine Learning (ICML), pages 1889–1897. PMLR, 2015.
  35. Forecasting exchange rate using deep belief networks and conjugate gradient method. Neurocomputing, 167:243–253, 2015. ISSN 0925-2312. https://doi.org/10.1016/j.neucom.2015.04.071. URL https://www.sciencedirect.com/science/article/pii/S0925231215005408.
  36. Semi-supervised learning using sparse eigenfunction bases. Advances in Neural Information Processing Systems (NeurIPS), 22, 2009.
  37. Applications of machine learning in drug discovery and development. Nature reviews Drug discovery, 18(6):463–477, 2019.
  38. Richard S Varga. Matrix iterative analysis. Prentice Hall Series in Automatic Computations, 1962.
  39. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261–272, 2020. 10.1038/s41592-019-0686-2.
  40. Nisheeth K Vishnoi. Laplacian solvers and their algorithmic applications. Theoretical Computer Science, 8(1-2):1–141, 2012.
  41. A class of gradient unconstrained minimization algorithms with adaptive stepsize. Journal of Computational and Applied Mathematics, 114(2):367–386, 2000. ISSN 0377-0427. https://doi.org/10.1016/S0377-0427(99)00276-9. URL https://www.sciencedirect.com/science/article/pii/S0377042799002769.
  42. Scalable semi-supervised learning by efficient anchor graph regularization. IEEE Transactions on Knowledge and Data Engineering, 28(7):1864–1877, 2016.
  43. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747, 2017.
  44. Xiaojin Zhu. Semi-supervised learning with graphs. Carnegie Mellon University, 2005.
  45. Learning from labeled and unlabeled data with label propagation. 2002.
  46. Introduction to semi-supervised learning. Synthesis lectures on artificial intelligence and machine learning, 3(1):1–130, 2009.
  47. Semi-supervised learning using Gaussian fields and harmonic functions. In International Conference on Machine Learning (ICML), pages 912–919, 2003.
Citations (1)

Summary

We haven't generated a summary for this paper yet.