Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Hopfield model with planted patterns: a teacher-student self-supervised learning model (2304.13710v3)

Published 26 Apr 2023 in cond-mat.dis-nn, cs.LG, math-ph, and math.MP

Abstract: While Hopfield networks are known as paradigmatic models for memory storage and retrieval, modern artificial intelligence systems mainly stand on the machine learning paradigm. We show that it is possible to formulate a teacher-student self-supervised learning problem with Boltzmann machines in terms of a suitable generalization of the Hopfield model with structured patterns, where the spin variables are the machine weights and patterns correspond to the training set's examples. We analyze the learning performance by studying the phase diagram in terms of the training set size, the dataset noise and the inference temperature (i.e. the weight regularization). With a small but informative dataset the machine can learn by memorization. With a noisy dataset, an extensive number of examples above a critical threshold is needed. In this regime the memory storage limits of the system becomes an opportunity for the occurrence of a learning regime in which the system can generalize.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (63)
  1. John J Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proceedings of the national academy of sciences, 79(8):2554–2558, 1982.
  2. Donald Olding Hebb. The organization of behavior: A neuropsychological theory. Psychology press, 2005.
  3. Statistical mechanics of neural networks near saturation. Annals of physics, 173(1):30–67, 1987.
  4. Theory of neural information processing systems. OUP Oxford, 2005.
  5. E Gardner. Multiconnected neural network models. Journal of Physics A: Mathematical and General, 20(11):3453, 1987.
  6. Multitasking associative networks. Physical review letters, 109(26):268101, 2012.
  7. Immune networks: multitasking capabilities near saturation. Journal of Physics A: Mathematical and Theoretical, 46(41):415003, 2013.
  8. Immune networks: multi-tasking capabilities at medium load. Journal of Physics A: Mathematical and Theoretical, 46(33):335101, 2013.
  9. Retrieving infinite numbers of patterns in a spin-glass model of immune networks. Europhysics Letters, 117(2):28003, 2017.
  10. Extensive parallel processing on scale-free networks. Physical review letters, 113(23):238106, 2014.
  11. Retrieval capabilities of hierarchical networks: From dyson to hopfield. Physical review letters, 114(2):028103, 2015.
  12. Hierarchical neural networks perform both serial and parallel processing. Neural Networks, 66:22–35, 2015.
  13. Metastable states in the hierarchical dyson model drive parallel processing in the hierarchical hopfield network. Journal of Physics A: Mathematical and Theoretical, 48(1):015001, 2014.
  14. Topological properties of hierarchical networks. Physical Review E, 91(6):062807, 2015.
  15. Phase diagram of restricted boltzmann machines and generalized hopfield networks with arbitrary priors. Physical Review E, 97(2):022310, 2018.
  16. How glassy are neural networks? Journal of Statistical Mechanics: Theory and Experiment, 2012(07):P07009, 2012.
  17. Neural networks retrieving boolean patterns in a sea of gaussian ones. Journal of Statistical Physics, 168:1085–1104, 2017.
  18. Legendre equivalences of spherical boltzmann machines. Journal of Physics A: Mathematical and Theoretical, 53(9):094001, 2020.
  19. Anergy in self-directed b lymphocytes: a statistical mechanics perspective. Journal of theoretical biology, 375:21–31, 2015.
  20. High storage capacity in the hopfield model with auto-interactions—stability analysis. Journal of Physics A: Mathematical and Theoretical, 50(46):465001, 2017.
  21. José Fernando Fontanari and WK Theumann. On the storage of correlated patterns in hopfield’s model. Journal de Physique, 51(5):375–386, 1990.
  22. Modified pseudo-inverse neural networks storing correlated patterns. Journal of Physics A: Mathematical and General, 25(10):2843, 1992.
  23. JL Van Hemmen. Hebbian learning, its correlation catastrophe, and unlearning. Network: Computation in Neural Systems, 8(3):V1, 1997.
  24. Storing, learning and retrieving biased patterns. Applied Mathematics and Computation, 415:126716, 2022.
  25. Matthias Löwe. On the storage capacity of hopfield models with correlated patterns. The Annals of Applied Probability, 8(4):1216–1250, 1998.
  26. Hanoch Gutfreund. Neural networks with hierarchically correlated patterns. Physical Review A, 37(2):570, 1988.
  27. Effect of spatial correlations on hopfield neural network and dense associative memories. Physica A: Statistical Mechanics and its Applications, 612:128487, 2023.
  28. The emergence of a concept in shallow neural networks. Neural Networks, 148:232–253, 2022.
  29. Supervised hebbian learning. Europhysics Letters, 2022.
  30. Unsupervised feature learning from finite data by message passing: discontinuous versus continuous phase transition. Physical Review E, 94(6):062310, 2016.
  31. Haiping Huang. Statistical mechanics of unsupervised feature learning in a restricted boltzmann machine with binary synapses. Journal of Statistical Mechanics: Theory and Experiment, 2017(5):053302, 2017.
  32. Marc Mézard. Mean-field message-passing equations in the hopfield model and its generalizations. Physical Review E, 95(2):022117, 2017.
  33. Statistical-mechanical study of deep boltzmann machine given weight parameters after training by singular value decomposition. Journal of the Physical Society of Japan, 91(11):114001, 2022.
  34. The hidden-manifold hopfield model and a learning phase transition. arXiv preprint arXiv:2303.16880, 2023.
  35. Self-supervised learning: Generative or contrastive. IEEE transactions on knowledge and data engineering, 35(1):857–876, 2021.
  36. Yukito Iba. The nishimori line and bayesian statistics. Journal of Physics A: Mathematical and General, 32(21):3875, 1999.
  37. The adaptive interpolation method: a simple scheme to prove replica formulas in bayesian inference. Probability theory and related fields, 174:1133–1185, 2019.
  38. Limits and performances of algorithms based on simulated annealing in solving sparse hard inference problems. Physical Review X, 13(2):021011, 2023.
  39. Mismatching as a tool to enhance algorithmic performances of monte carlo methods for the planted clique model. Journal of Statistical Mechanics: Theory and Experiment, 2021(11):113406, 2021.
  40. A learning algorithm for boltzmann machines. Cognitive science, 9(1):147–169, 1985.
  41. Multi-species mean field spin glasses. rigorous results. In Annales Henri Poincaré, volume 16, pages 691–708. Springer, 2015.
  42. Dmitry Panchenko. The free energy in a multi-species sherrington–kirkpatrick model. 2015.
  43. The multi-species mean-field spin-glass on the nishimori line. Journal of Statistical Physics, 182:1–20, 2021.
  44. Annealing and replica-symmetry in deep boltzmann machines. Journal of Statistical Physics, 180:665–677, 2020.
  45. Deep boltzmann machines: rigorous results at arbitrary depth. In Annales Henri Poincaré, volume 22, pages 2619–2642. Springer, 2021.
  46. Giuseppe Genovese. A remark on the spherical bipartite spin glass. Mathematical Physics, Analysis and Geometry, 25(2):14, 2022.
  47. Non-convex multipartite ferromagnets. Journal of Statistical Physics, 163:492–513, 2016.
  48. Legendre duality of spherical and gaussian spin glasses. Mathematical Physics, Analysis and Geometry, 18:1–19, 2015.
  49. Mean field bipartite spin models treated with mechanical techniques. The European Physical Journal B, 87:1–13, 2014.
  50. Overlap synchronisation in multipartite random energy models. Journal of Statistical Physics, 169:1162–1170, 2017.
  51. About a solvable mean field model of a gaussian spin glass. Journal of Physics A: Mathematical and Theoretical, 47(15):155002, 2014.
  52. On the equivalence of hopfield networks and boltzmann machines. Neural Networks, 34:1–9, 2012.
  53. Non-convex multi-species hopfield models. Journal of Statistical Physics, 172(5):1247–1269, 2018.
  54. Phase transitions in restricted boltzmann machines with generic priors. Physical Review E, 96(4):042156, 2017.
  55. High-dimensional inference with the generalized hopfield model: Principal component analysis and corrections. Physical Review E, 83(5):051123, 2011.
  56. Minimal model of permutation symmetry in unsupervised learning. Journal of Physics A: Mathematical and Theoretical, 52(41):414001, 2019.
  57. Inverse problems for structured datasets using parallel tap equations and restricted boltzmann machines. Scientific Reports, 11(1):19990, 2021.
  58. Inference and learning in sparse systems with multiple states. Physical Review E, 83(5):056114, 2011.
  59. Solving the inverse ising problem by mean-field methods in a clustered phase space with many states. Physical Review E, 94(1):012112, 2016.
  60. Hidetoshi Nishimori. Statistical physics of spin glasses and information processing: an introduction. Number 111. Clarendon Press, 2001.
  61. The solution of the deep boltzmann machine on the nishimori line. Communications in Mathematical Physics, 387:1191–1214, 2021.
  62. Information, physics, and computation. Oxford University Press, 2009.
  63. Richard S Ellis. Entropy, large deviations, and statistical mechanics, volume 1431. Taylor & Francis, 2006.
Citations (8)

Summary

We haven't generated a summary for this paper yet.