Papers
Topics
Authors
Recent
2000 character limit reached

Pretraining with Random Noise for Fast and Robust Learning without Weight Transport (2405.16731v2)

Published 27 May 2024 in cs.LG and cs.NE

Abstract: The brain prepares for learning even before interacting with the environment, by refining and optimizing its structures through spontaneous neural activity that resembles random noise. However, the mechanism of such a process has yet to be thoroughly understood, and it is unclear whether this process can benefit the algorithm of machine learning. Here, we study this issue using a neural network with a feedback alignment algorithm, demonstrating that pretraining neural networks with random noise increases the learning efficiency as well as generalization abilities without weight transport. First, we found that random noise training modifies forward weights to match backward synaptic feedback, which is necessary for teaching errors by feedback alignment. As a result, a network with pre-aligned weights learns notably faster than a network without random noise training, even reaching a convergence speed comparable to that of a backpropagation algorithm. Sequential training with both random noise and data brings weights closer to synaptic feedback than training solely with data, enabling more precise credit assignment and faster learning. We also found that each readout probability approaches the chance level and that the effective dimensionality of weights decreases in a network pretrained with random noise. This pre-regularization allows the network to learn simple solutions of a low rank, reducing the generalization loss during subsequent training. This also enables the network robustly to generalize a novel, out-of-distribution dataset. Lastly, we confirmed that random noise pretraining reduces the amount of meta-loss, enhancing the network ability to adapt to various tasks. Overall, our results suggest that random noise training with feedback alignment offers a straightforward yet effective method of pretraining that facilitates quick and reliable learning without weight transport.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (78)
  1. Prenatal tuning of thalamic spontaneous activity patterns regulates somatosensory map resolution. bioRxiv, pages 2024–03, 2024.
  2. Innate visual learning through spontaneous activity patterns. PLoS Computational Biology, 4(8):e1000137, 2008.
  3. Spontaneous activity in the zebrafish tectum reorganizes over development and is influenced by visual experience. Current Biology, 27(16):2407–2419, 2017.
  4. Spontaneous neuronal activity in developing neocortical networks: from single cells to large-scale interactions. Frontiers in neural circuits, 10:40, 2016.
  5. Richard H Masland. Maturation of function in the developing rabbit retina. Journal of Comparative Neurology, 175(3):275–286, 1977.
  6. Spontaneous impulse activity of rat retinal ganglion cells in prenatal life. Science, 242(4875):90–91, 1988.
  7. Retinal waves coordinate patterned activity throughout the developing visual system. Nature, 490(7419):219–225, 2012.
  8. Development of activity in the mouse visual cortex. Journal of Neuroscience, 36(48):12259–12275, 2016.
  9. Activity-dependent development of visual receptive fields. Current opinion in neurobiology, 42:136–143, 2017.
  10. Role of emergent neural activity in visual map development. Current opinion in neurobiology, 24:166–175, 2014.
  11. Developmentally regulated spontaneous activity in the embryonic chick retina. Journal of Neuroscience, 18(21):8839–8852, 1998.
  12. Developmental spontaneous activity promotes formation of sensory domains, frequency tuning and proper gain in central auditory circuits. Cell reports, 41(7), 2022.
  13. Development of hearing in neonatal rats: air and bone conducted abr thresholds. Hearing research, 69(1-2):236–242, 1993.
  14. Transient subgranular hyperconnectivity to l2/3 and enhanced pairwise correlations during the critical period in the mouse auditory cortex. Cerebral Cortex, 30(3):1914–1930, 2020.
  15. Helen J Kennedy. New developments in understanding the mechanisms and function of spontaneous electrical activity in the developing mammalian auditory system. Journal of the Association for Research in Otolaryngology, 13:437–445, 2012.
  16. The wiring of developing sensory circuits—from patterned spontaneous activity to synaptic plasticity mechanisms. Frontiers in neural circuits, 10:71, 2016.
  17. Spontaneous activity in the developing auditory system. Cell and tissue research, 361:65–75, 2015.
  18. Early postnatal development of spontaneous and acoustically evoked discharge activity of principal cells of the medial nucleus of the trapezoid body: an in vivo study in mice. Journal of Neuroscience, 29(30):9510–9520, 2009.
  19. Spontaneous motor activity in fetal and infant rats is organized into discrete multilimb bouts. Behavioral neuroscience, 114(2):328, 2000.
  20. Viktor Hamburger and CH Narayanan. Effects of the deafferentation of the trigeminal area on the motility of the chick embryo. Journal of Experimental Zoology, 170(4):411–426, 1969.
  21. Motility in the chick embryo in the absence of sensory input. Journal of Experimental Zoology, 162(2):133–159, 1966.
  22. Electrical activity in the spinal cord of the chick embryo, in situ. Proceedings of the National Academy of Sciences, 65(3):508–515, 1970.
  23. Sensory feedback synchronizes motor and sensory neuronal networks in the neonatal rat spinal cord. Nature communications, 7(1):13060, 2016.
  24. Development of tactile sensory circuits in the cns. Current opinion in neurobiology, 53:66–75, 2018.
  25. Prenatal activity from thalamic neurons governs the emergence of functional cortical maps in mice. Science, 364(6444):987–990, 2019.
  26. Patchwork-type spontaneous activity in neonatal barrel cortex layer 4 transmitted via thalamocortical projections. Cell reports, 22(1):123–135, 2018.
  27. Spontaneous activity in developing thalamic and cortical sensory networks. Neuron, 109(16):2519–2534, 2021.
  28. Electrical activity controls area-specific expression of neuronal apoptosis in the mouse developing cerebral cortex. Elife, 6:e27696, 2017.
  29. Synaptic activity and the construction of cortical circuits. Science, 274(5290):1133–1138, 1996.
  30. Electrical activity patterns and the functional maturation of the neocortex. European Journal of Neuroscience, 34(10):1677–1686, 2011.
  31. Shaping brain connections through spontaneous neural activity. European Journal of Neuroscience, 35(10):1595–1604, 2012.
  32. Visual number sense in untrained deep neural networks. Science advances, 7(1):eabd6127, 2021.
  33. Face detection in untrained deep neural networks. Nature communications, 12(1):7328, 2021.
  34. Invariance of object detection in untrained deep neural networks. Frontiers in Computational Neuroscience, 16:1030707, 2022.
  35. Comparison of visual quantities in untrained neural networks. Cell Reports, 42(8), 2023.
  36. Backpropagation and the brain. Nature Reviews Neuroscience, 21(6):335–346, 2020.
  37. Learning representations by back-propagating errors. Nature, 323:533–536, 10 1986.
  38. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 2012.
  39. Deep learning. nature, 521(7553):436–444, 2015.
  40. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. arXiv preprint arXiv:1312.6120, 2013.
  41. Supervised and unsupervised learning with two sites of synaptic integration. Journal of computational neuroscience, 11:207–215, 2001.
  42. Stephen Grossberg. Competitive learning: From interactive activation to adaptive resonance. Cognitive science, 11(1):23–63, 1987.
  43. Francis Crick. The recent excitement about neural networks. Nature, 337(6203):129–132, 1989.
  44. Random synaptic feedback weights support error backpropagation for deep learning. Nature communications, 7(1):13276, 2016.
  45. Deep learning without weight transport. Advances in neural information processing systems, 32, 2019.
  46. Arild Nøkland. Direct feedback alignment provides learning in deep neural networks. Advances in neural information processing systems, 29, 2016.
  47. Mother-stranger face discrimination by the newborn. Infant Behavior and development, 7(1):19–25, 1984.
  48. Neonatal recognition of the mother’s face. British journal of developmental psychology, 7(1):3–15, 1989.
  49. Mother’s face recognition by neonates: A replication and an extension. Infant behavior and development, 18(1):79–85, 1995.
  50. Traer Scott. Wild Babies: Photographs of Baby Animals from Giraffes to Hummingbirds. Chronicle Books, 2016.
  51. Irving John Good. Rational decisions. Journal of the Royal Statistical Society: Series B (Methodological), 14(1):107–114, 1952.
  52. Li Deng. The mnist database of handwritten digit images for machine learning research. IEEE signal processing magazine, 29(6):141–142, 2012.
  53. Principal component analysis. Chemometrics and Intelligent Laboratory Systems, 2:37–52, 8 1987.
  54. The effective rank: A measure of effective dimensionality. In 2007 15th European signal processing conference, pages 606–610. IEEE, 2007.
  55. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747, 2017.
  56. Deep learning for classical japanese literature. arXiv preprint arXiv:1812.01718, 2018.
  57. Model-agnostic meta-learning for fast adaptation of deep networks. In International conference on machine learning, pages 1126–1135. PMLR, 2017.
  58. Donald Olding Hebb. The organization of behavior: A neuropsychological theory. Psychology press, 1949.
  59. Spike-timing-dependent synaptic modification induced by natural spike trains. Nature, 416(6879):433–438, 2002.
  60. Spike timing-dependent plasticity of neural circuits. Neuron, 44(1):23–30, 2004.
  61. Spike timing–dependent plasticity: a hebbian learning rule. Annu. Rev. Neurosci., 31:25–46, 2008.
  62. H Sebastian Seung. Learning in spiking neural networks by reinforcement of stochastic synaptic transmission. Neuron, 40(6):1063–1073, 2003.
  63. An unsupervised stdp-based spiking neural network inspired by biologically plausible learning rules and connections. Neural Networks, 165:799–808, 2023.
  64. Implicit regularization in deep matrix factorization. Advances in Neural Information Processing Systems, 32, 2019.
  65. The low-rank simplicity bias in deep networks. arXiv preprint arXiv:2103.10427, 2021.
  66. Implicit regularization via neural feature alignment. In International Conference on Artificial Intelligence and Statistics, pages 2269–2277. PMLR, 2021.
  67. Dan Zhao. Combining explicit and implicit regularization for efficient learning in deep networks. Advances in Neural Information Processing Systems, 35:3024–3038, 2022.
  68. Representation based complexity measures for predicting generalization in deep learning. arXiv preprint arXiv:2012.02775, 2020.
  69. The information bottleneck method. arXiv preprint physics/0004057, 2000.
  70. Deep learning and the information bottleneck principle. In 2015 ieee information theory workshop (itw), pages 1–5. IEEE, 2015.
  71. How does information bottleneck help deep learning? In International Conference on Machine Learning, pages 16049–16096. PMLR, 2023.
  72. Opening the black box of deep neural networks via information. arXiv preprint arXiv:1703.00810, 2017.
  73. On the information bottleneck theory of deep learning. Journal of Statistical Mechanics: Theory and Experiment, 2019(12):124020, 2019.
  74. Visualizing the loss landscape of neural nets. Advances in neural information processing systems, 31, 2018.
  75. Loss landscapes are all you need: Neural network generalization can be explained without the implicit bias of gradient descent. In The Eleventh International Conference on Learning Representations, 2022.
  76. How does loss function affect generalization performance of deep learning? application to human age estimation. In International Conference on Machine Learning, pages 141–151. PMLR, 2021.
  77. Learning multiple layers of features from tiny images. 2009.
  78. Jonathan J. Hull. A database for handwritten text recognition research. IEEE Transactions on pattern analysis and machine intelligence, 16(5):550–554, 1994.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 1 like about this paper.