Cooling schedules for learning in neural networks
- 1 June 1993
- journal article
- research article
- Published by American Physical Society (APS) in Physical Review E
- Vol. 47 (6) , 4457-4464
- https://doi.org/10.1103/physreve.47.4457
Abstract
We derive cooling schedules for the global optimization of learning in neural networks. We discuss a two-level system with one global and one local minimum. The analysis is extended to systems with many minima. The optimal cooling schedule is (asymptotically) of the form η(t)=/lnt, with η(t) the learning parameter at time t and a constant, dependent on the reference learning parameters for the various transitions. In some simple cases, can be calculated. Simulations confirm the theoretical results.
Keywords
This publication has 14 references indexed in Scilit:
- Learning in neural networks with local minimaPhysical Review A, 1992
- Learning processes in neural networksPhysical Review A, 1991
- Convergence of learning algorithms with constant learning ratesIEEE Transactions on Neural Networks, 1991
- Convergence properties of Kohonen's topology conserving maps: fluctuations, stability, and dimension selectionBiological Cybernetics, 1988
- Asymptotic Global Behavior for Stochastic Approximation and Diffusions with Slowly Decreasing Noise Effects: Global Minimization via Monte CarloSIAM Journal on Applied Mathematics, 1987
- Learning representations by back-propagating errorsNature, 1986
- Robustness and Approximation of Escape Times and Large Deviations Estimates for Systems with Small Noise EffectsSIAM Journal on Applied Mathematics, 1984
- Optimization by Simulated AnnealingScience, 1983
- Self-organized formation of topologically correct feature mapsBiological Cybernetics, 1982
- Analysis of recursive stochastic algorithmsIEEE Transactions on Automatic Control, 1977