Learning in neural networks by reinforcement of irregular spiking
- 30 April 2004
- journal article
- research article
- Published by American Physical Society (APS) in Physical Review E
- Vol. 69 (4) , 041909
- https://doi.org/10.1103/physreve.69.041909
Abstract
Artificial neural networks are often trained by using the back propagation algorithm to compute the gradient of an objective function with respect to the synaptic strengths. For a biological neural network, such a gradient computation would be difficult to implement, because of the complex dynamics of intrinsic and synaptic conductances in neurons. Here we show that irregular spiking similar to that observed in biological neurons could be used as the basis for a learning rule that calculates a stochastic approximation to the gradient. The learning rule is derived based on a special class of model networks in which neurons fire spike trains with Poisson statistics. The learning is compatible with forms of synaptic dynamics such as short-term facilitation and depression. By correlating the fluctuations in irregular spiking with a reward signal, the learning rule performs stochastic gradient ascent on the expected reward. It is applied to two examples, learning the XOR computation and learning direction selectivity using depressing synapses. We also show in simulation that the learning rule is applicable to a network of noisy integrate-and-fire neurons. DOI: http://dx.doi.org/10.1103/PhysRevE.69.041909 © 2004 The American Physical SocietyKeywords
This publication has 20 references indexed in Scilit:
- Emergent SynthesisArtificial Intelligence in Engineering, 2001
- Efficient temporal processing with biologically realistic dynamic synapsesNetwork: Computation in Neural Systems, 2001
- Synaptic plasticity in a cerebellum-like structure depends on temporal orderNature, 1997
- Regulation of Synaptic Efficacy by Coincidence of Postsynaptic APs and EPSPsScience, 1997
- Synaptic Depression and Cortical Gain ControlScience, 1997
- Chaos in Neuronal Networks with Balanced Excitatory and Inhibitory ActivityScience, 1996
- Temporal Information Transformed into a Spatial Code by a Neural Network with Realistic PropertiesScience, 1995
- Weight perturbation: an optimal architecture and learning technique for analog VLSI feedforward and recurrent multilayer networksIEEE Transactions on Neural Networks, 1992
- Short-Term Synaptic PlasticityAnnual Review of Neuroscience, 1989
- Learning representations by back-propagating errorsNature, 1986