Inferring effective couplings with Restricted Boltzmann Machines
Abstract: Generative models offer a direct way of modeling complex data. Energy-based models attempt to encode the statistical correlations observed in the data at the level of the Boltzmann weight associated with an energy function in the form of a neural network. We address here the challenge of understanding the physical interpretation of such models. In this study, we propose a simple solution by implementing a direct mapping between the Restricted Boltzmann Machine and an effective Ising spin Hamiltonian. This mapping includes interactions of all possible orders, going beyond the conventional pairwise interactions typically considered in the inverse Ising (or Boltzmann Machine) approach, and allowing the description of complex datasets. Earlier works attempted to achieve this goal, but the proposed mappings were inaccurate for inference applications, did not properly treat the complexity of the problem, or did not provide precise prescriptions for practical application. To validate our method, we performed several controlled inverse numerical experiments in which we trained the RBMs using equilibrium samples of predefined models with local external fields, 2-body and 3-body interactions in different sparse topologies. The results demonstrate the effectiveness of our proposed approach in learning the correct interaction network and pave the way for its application in modeling interesting binary variable datasets. We also evaluate the quality of the inferred model based on different training methods.
- D. H. Ackley, G. E. Hinton and T. J. Sejnowski, A learning algorithm for Boltzmann machines, Cognitive science 9(1), 147 (1985).
- P. Smolensky, In Parallel Distributed Processing: Volume 1 by D. Rumelhart and J. McLelland, chap. 6: Information Processing in Dynamical Systems: Foundations of Harmony Theory, 194-281. MIT Press (1986).
- Filters, random fields and maximum entropy (frame): Towards a unified theory for texture modeling, International Journal of Computer Vision 27, 107 (1998).
- A tutorial on energy-based learning, Predicting structured data 1(0) (2006).
- A theory of generative convnet, In International Conference on Machine Learning, pp. 2635–2644. PMLR (2016).
- Inverse statistical problems: from the inverse Ising problem to data science, Advances in Physics 66(3), 197 (2017).
- Searching for collective behavior in a large network of sensory neurons, PLOS Computational Biology 10(1), 1 (2014), 10.1371/journal.pcbi.1003408.
- Direct-coupling analysis of residue coevolution captures native contacts across many protein families, Proceedings of the National Academy of Sciences 108(49), E1293 (2011), 10.1073/pnas.1111471108, https://www.pnas.org/doi/pdf/10.1073/pnas.1111471108.
- H.-L. Zeng and E. Aurell, Inferring genetic fitness from genomic data, Physical Review E 101(5), 052409 (2020).
- Statistical genetics and evolution of quantitative traits, Reviews of Modern Physics 83(4), 1283 (2011).
- Using the principle of entropy maximization to infer genetic interaction networks from gene expression patterns, Proceedings of the National Academy of Sciences 103(50), 19033 (2006), 10.1073/pnas.0609152103, https://www.pnas.org/doi/pdf/10.1073/pnas.0609152103.
- Gradient-based learning applied to document recognition, Proceedings of the IEEE 86(11), 2278 (1998).
- Optimal perceptual inference, In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, vol. 448, pp. 448–453. Citeseer (1983).
- L. Deng, The mnist database of handwritten digit images for machine learning research, IEEE Signal Processing Magazine 29(6), 141 (2012).
- Machine learning determination of dynamical parameters: The Ising model case, Phys. Rev. B 100, 064304 (2019), 10.1103/PhysRevB.100.064304.
- N. Bulso and Y. Roudi, Restricted Boltzmann machines as models of interacting variables, Neural Computation 33(10), 2646 (2021).
- A. Decelle and C. Furtlehner, Restricted Boltzmann machine: Recent advances and mean-field theory*, Chinese Physics B 30(4), 040202 (2021), 10.1088/1674-1056/abd160.
- Phase diagram of restricted Boltzmann machines and generalized Hopfield networks with arbitrary priors, Physical Review E 97(2), 022310 (2018).
- J. Tubiana and R. Monasson, Emergence of compositional representations in restricted Boltzmann machines, Physical review letters 118(13), 138301 (2017).
- A. Decelle, G. Fissore and C. Furtlehner, Thermodynamics of restricted Boltzmann machines and related learning dynamics, Journal of Statistical Physics 172, 1576 (2018).
- A. Decelle, G. Fissore and C. Furtlehner, Spectral dynamics of learning in restricted Boltzmann machines, Europhysics Letters 119(6), 60001 (2017).
- Training restricted Boltzmann machine via the Thouless-Anderson-Palmer free energy, Advances in neural information processing systems 28 (2015).
- Deterministic and generalized framework for unsupervised learning with restricted Boltzmann machines, Physical Review X 8(4), 041006 (2018).
- A. Decelle, B. Seoane and L. Rosset, Unsupervised hierarchical clustering using the learning dynamics of restricted Boltzmann machines, Physical Review E 108(1), 014110 (2023).
- On the equivalence of Hopfield networks and Boltzmann machines, Neural Networks 34, 1 (2012), https://doi.org/10.1016/j.neunet.2012.06.003.
- S. Feng, D. Kong and N. Trivedi, A statistical approach to topological entanglement: Boltzmann machine representation of higher-order irreducible correlation, arXiv preprint arXiv:2302.03212 (2023).
- E. Rrapaj and A. Roggero, Exact representations of many-body interactions with restricted-boltzmann-machine neural networks, Physical Review E 103(1), 013302 (2021).
- A density consistency approach to the inverse ising problem, Journal of Statistical Mechanics: Theory and Experiment 2021(3), 033416 (2021), 10.1088/1742-5468/abed43.
- A. d. J. Navas Gómez, Exploring in and out-of-equilibrium learning regimes of restricted Boltzmann machines, Master’s thesis, Universidad Nacional de Colombia (2022).
- M. Mézard and T. Mora, Constraint satisfaction problems and neural networks: A statistical physics perspective, Journal of Physiology-Paris 103(1), 107 (2009), https://doi.org/10.1016/j.jphysparis.2009.05.013, Neuromathematics of Vision.
- H. C. Nguyen and J. Berg, Bethe–Peierls approximation and the inverse Ising problem, Journal of Statistical Mechanics: Theory and Experiment 2012(03), P03004 (2012), 10.1088/1742-5468/2012/03/P03004.
- F. Ricci-Tersenghi, The Bethe approximation for solving the inverse Ising problem: a comparison with other inference methods, Journal of Statistical Mechanics: Theory and Experiment 2012(08), P08015 (2012), 10.1088/1742-5468/2012/08/P08015.
- O. Krause, A. Fischer and C. Igel, Algorithms for estimating the partition function of restricted boltzmann machines, Artificial Intelligence 278, 103195 (2020).
- R. M. Neal, Annealed importance sampling, Statistics and computing 11, 125 (2001).
- A. Decelle, C. Furtlehner and B. Seoane, Equilibrium and non-equilibrium regimes in the learning of restricted Boltzmann machines, In M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang and J. W. Vaughan, eds., Advances in Neural Information Processing Systems, vol. 34, pp. 5345–5359. Curran Associates, Inc. (2021).
- Explaining the effects of non-convergent sampling in the training of energy-based models, arXiv preprint arXiv:2301.09428 (2023).
- Learning a restricted Boltzmann machine using biased Monte Carlo sampling, SciPost Phys. 14, 032 (2023), 10.21468/SciPostPhys.14.3.032.
- Theory of spin glasses, Journal of Physics F: Metal Physics 5(5), 965 (1975).
- Fast and functional structured data generators rooted in out-of-equilibrium physics, arXiv preprint arXiv:2307.06797 (2023).
- Creating artificial human genomes using generative neural networks, PLoS genetics 17(2), e1009303 (2021).
- Deep convolutional and conditional neural networks for large-scale genomic data generation, PLOS Computational Biology 19(10), 1 (2023), 10.1371/journal.pcbi.1011584.
- E. Mauri, S. Cocco and R. Monasson, Mutational paths with sequence-based models of proteins: from sampling to mean-field characterization, Physical Review Letters 130(15), 158402 (2023).
- A. Fischer and C. Igel, An introduction to restricted Boltzmann machines, In L. Alvarez, M. Mejail, L. Gomez and J. Jacobo, eds., Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications, pp. 14–36. Springer Berlin Heidelberg, Berlin, Heidelberg, ISBN 978-3-642-33275-3, https://doi.org/10.1007/978-3-642-33275-3_2 (2012).
- T. Tieleman, Training restricted Boltzmann machines using approximations to the likelihood gradient, ICML ’08, p. 1064–1071. Association for Computing Machinery, New York, NY, USA, ISBN 9781605582054, 10.1145/1390156.1390290 (2008).
- Learning non-convergent non-persistent short-run mcmc toward energy-based model, In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox and R. Garnett, eds., Advances in Neural Information Processing Systems, vol. 32. Curran Associates, Inc. (2019).
- Theory of neural information processing systems, OUP Oxford (2005).
- H. Touchette, A basic introduction to large deviations: Theory, applications, simulations, arXiv preprint arXiv:1106.4146 (2011).
- W. K. Hastings, Monte carlo sampling methods using markov chains and their applications, Biometrika 57(1), 97 (1953).
- Equation of state calculations by fast computing machines, The journal of chemical physics 21(6), 1087 (1953).
- U. Wolff, Collective monte carlo updating for spin systems, Physical Review Letters 62(4), 361 (1989).
- D. J. Amit and V. Martin-Mayor, Field theory, the renormalization group, and critical phenomena: graphs to computers, World Scientific Publishing Company (2005).
- H. Ballesteros and V. Martín-Mayor, Test for random number generators: Schwinger-dyson equations for the ising model, Physical Review E 58(5), 6787 (1998).
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.