On the role of synaptic stochasticity in training low-precision neural networks (1710.09825v2)
Abstract: Stochasticity and limited precision of synaptic weights in neural network models are key aspects of both biological and hardware modeling of learning processes. Here we show that a neural network model with stochastic binary weights naturally gives prominence to exponentially rare dense regions of solutions with a number of desirable properties such as robustness and good generalization performance, while typical solutions are isolated and hard to find. Binary solutions of the standard perceptron problem are obtained from a simple gradient descent procedure on a set of real values parametrizing a probability distribution over the binary synapses. Both analytical and numerical results are presented. An algorithmic extension aimed at training discrete deep neural networks is also investigated.
- Carlo Baldassi (36 papers)
- Federica Gerace (13 papers)
- Hilbert J. Kappen (22 papers)
- Carlo Lucibello (38 papers)
- Luca Saglietti (21 papers)
- Enzo Tartaglione (68 papers)
- Riccardo Zecchina (48 papers)