A Theory of Synaptic Neural Balance: From Local to Global Order
Abstract: We develop a general theory of synaptic neural balance and how it can emerge or be enforced in neural networks. For a given regularizer, a neuron is said to be in balance if the total cost of its input weights is equal to the total cost of its output weights. The basic example is provided by feedforward networks of ReLU units trained with $L_2$ regularizers, which exhibit balance after proper training. The theory explains this phenomenon and extends it in several directions. The first direction is the extension to bilinear and other activation functions. The second direction is the extension to more general regularizers, including all $L_p$ regularizers. The third direction is the extension to non-layered architectures, recurrent architectures, convolutional architectures, as well as architectures with mixed activation functions. Gradient descent on the error function alone does not converge in general to a balanced state, where every neuron is in balance, even when starting from a balanced state. However, gradient descent on the regularized error function ought to converge to a balanced state, and thus network balance can be used to assess learning progress. The theory is based on two local neuronal operations: scaling which is commutative, and balancing which is not commutative. Given any initial set of weights, when local balancing operations are applied to each neuron in a stochastic manner, global order always emerges through the convergence of the stochastic balancing algorithm to the same unique set of balanced weights. The reason for this is the existence of an underlying strictly convex optimization problem where the relevant variables are constrained to a linear, only architecture-dependent, manifold. Simulations show that balancing neurons prior to learning, or during learning in alternation with gradient descent steps, can improve learning speed and final performance.
- P. Baldi. Autoencoders, Unsupervised Learning, and Deep Architectures. Journal of Machine Learning Research. Proceedings of 2011 ICML Workshop on Unsupervised and Transfer Learning, 27:37–50, 2012.
- P. Baldi. Deep Learning in Science. Cambridge University Press, Cambridge, UK, 2021.
- P. Baldi and K. Hornik. Neural networks and principal component analysis: Learning from examples without local minima. Neural Networks, 2(1):53–58, 1989.
- Pierre Baldi. Deep learning over-parameterization: the shallow fallacy. In Northern Lights Deep Learning Conference, pages 7–12. Proceedings Machine Learning Research, 2024.
- A theory of local learning, the learning channel, and the optimality of backpropagation. Neural Networks, 83:61–74, 2016.
- The capacity of feedforward neural networks. Neural Networks, 116:288–311, 2019. Also: arXiv preprint arXiv:1901.00434.
- Homeostatic role of heterosynaptic plasticity: models and experiments. Frontiers in computational neuroscience, 9:89, 2015.
- Li Deng. The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141–142, 2012.
- Algorithmic regularization in learning deep homogeneous models: Layers are automatically balanced. Advances in Neural Information Processing Systems, 31, 2018.
- Heterosynaptic plasticity determines the set point for cortical excitatory-inhibitory balance. Neuron, 106(5):842–854, 2020.
- Spike-threshold adaptation predicted by membrane potential dynamics in vivo. PLoS computational biology, 10(4):e1003560, 2014.
- Robert C Froemke. Plasticity of cortical excitatory-inhibitory balance. Annual review of neuroscience, 38:195–219, 2015.
- Integrating the neurodevelopmental and dopamine hypotheses of schizophrenia and the role of cortical excitation-inhibition balance. Biological psychiatry, 2022.
- Adaptive spike threshold enables robust and temporally precise neuronal encoding. PLoS computational biology, 12(6):e1004984, 2016.
- Neuromorphic artificial intelligence systems. Frontiers in Neuroscience, 16:1513, 2022.
- Neutrams: Neural network transformation and co-design under neuromorphic hardware constraints. In 2016 49th Annual IEEE/ACM International Symposium on Microarchitecture (MICRO), pages 1–13. IEEE, 2016.
- Neurotransmitter-induced excitatory and inhibitory functions in artificial synapses. Advanced Functional Materials, 32(21):2200497, 2022.
- Learning multiple layers of features from tiny images. 2009.
- Data-dependent path normalization in neural networks. arXiv preprint arXiv:1511.06747, 2015.
- Optimal approximation of piecewise smooth functions using deep relu neural networks. Neural Networks, 108:296–330, 2018.
- Model of autism: increased ratio of excitation/inhibition in key neural systems. Genes, Brain and Behavior, 2(5):255–267, 2003.
- Conversion of continuous-valued deep networks to efficient event-driven networks for image classification. Frontiers in neuroscience, 11:294078, 2017.
- On the physiological and structural contributors to the dynamic balance of excitation and inhibition in local cortical networks. bioRxiv, pages 2023–01, 2023.
- Optimizing the energy consumption of spiking neural networks for neuromorphic applications. Frontiers in neuroscience, 14:662, 2020.
- Synaptic balancing: A biologically plausible local learning rule that provably increases neural network noise robustness without sacrificing task performance. PLOS Computational Biology, 18(9):e1010418, 2022.
- Neurophysiology and regulation of the balance between excitation and inhibition in neocortical circuits. Biological psychiatry, 81(10):821–831, 2017.
- SPLASH: Learnable activation functions for improving accuracy and adversarial robustness. Neural Networks, 140:1–12, 2021. Also: arXiv:2006.08947.
- Gina Turrigiano. Homeostatic synaptic plasticity: local and global mechanisms for stabilizing neuronal function. Cold Spring Harbor perspectives in biology, 4(1):a005736, 2012.
- Gina G Turrigiano. The dialectic of hebb and homeostasis. Philosophical transactions of the royal society B: biological sciences, 372(1715):20160258, 2017.
- Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science, 274(5293):1724–1726, 1996.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.