Efficient Learning in Boltzmann Machines Using Linear Response Theory
- 1 July 1998
- journal article
- research article
- Published by MIT Press in Neural Computation
- Vol. 10 (5) , 1137-1156
- https://doi.org/10.1162/089976698300017386
Abstract
The learning process in Boltzmann machines is computationally very expensive. The computational complexity of the exact algorithm is exponential in the number of neurons. We present a new approximate learning algorithm for Boltzmann machines, based on mean-field theory and the linear response theorem. The computational complexity of the algorithm is cubic in the number of neurons.In the absence of hidden units, we show how the weights can be directly computed from the fixed-point equation of the learning rules. Thus, in this case we do not need to use a gradient descent procedure for the learning process. We show that the solutions of this method are close to the optimal solutions and give a significant improvement when correlations play a significant role. Finally, we apply the method to a pattern completion task and show good performance for networks up to 100 neurons.This publication has 17 references indexed in Scilit:
- Stimulus-dependent correlations in stochastic networksPhysical Review E, 1997
- The Helmholtz MachineNeural Computation, 1995
- The invisible hand algorithm: Solving the assignment problem with statistical physicsNeural Networks, 1994
- Learning in Boltzmann TreesNeural Computation, 1994
- Theory of correlations in stochastic neural networksPhysical Review E, 1994
- The limitations of deterministic Boltzmann machine learningNetwork: Computation in Neural Systems, 1993
- Stereo integration, mean field theory and psychophysicsNetwork: Computation in Neural Systems, 1991
- Direct Determination of the Probability Distribution for the Spin-Glass Order ParameterPhysical Review Letters, 1983
- Solution of 'Solvable model of a spin glass'Philosophical Magazine, 1977
- Electric Moments of Molecules in LiquidsJournal of the American Chemical Society, 1936