Natural Gradient Descent for On-Line Learning
- 14 December 1998
- journal article
- research article
- Published by American Physical Society (APS) in Physical Review Letters
- Vol. 81 (24) , 5461-5464
- https://doi.org/10.1103/physrevlett.81.5461
Abstract
Natural gradient descent is an on-line variable-metric optimization algorithm which utilizes an underlying Riemannian parameter space. We analyze the dynamics of natural gradient descent beyond the asymptotic regime by employing an exact statistical mechanics description of learning in two-layer feed-forward neural networks. For a realizable learning scenario we find significant improvements over standard gradient descent for both the transient and asymptotic stages of learning, with a slower power law increase in learning time as task complexity grows.Keywords
This publication has 8 references indexed in Scilit:
- Natural Gradient Works Efficiently in LearningNeural Computation, 1998
- Globally Optimal Parameters for On-Line Learning in Multilayer Neural NetworksPhysical Review Letters, 1997
- On-line learning with adaptive back-propagation in two-layer networksPhysical Review E, 1997
- Transient dynamics of on-line learning in two-layered neural networksJournal of Physics A: General Physics, 1996
- On-line learning in soft committee machinesPhysical Review E, 1995
- Exact Solution for On-Line Learning in Multilayer Neural NetworksPhysical Review Letters, 1995
- Statistical Theory of Learning Curves under Entropic Loss CriterionNeural Computation, 1993
- Approximation by superpositions of a sigmoidal functionMathematics of Control, Signals, and Systems, 1989