A theoretical comparison of batch-mode, on-line, cyclic, and almost-cyclic learning
- 1 July 1996
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE Transactions on Neural Networks
- Vol. 7 (4) , 919-925
- https://doi.org/10.1109/72.508935
Abstract
We study and compare different neural network learning strategies: batch-mode learning, online learning, cyclic learning, and almost-cyclic learning. Incremental learning strategies require less storage capacity than batch-mode learning. However, due to the arbitrariness in the presentation order of the training patterns, incremental learning is a stochastic process; whereas batch-mode learning is deterministic. In zeroth order, i.e., as the learning parameter /spl eta/ tends to zero, all learning strategies approximate the same ordinary differential equation for convenience referred to as the "ideal behavior". Using stochastic methods valid for small learning parameters /spl eta/, we derive differential equations describing the evolution of the lowest-order deviations from this ideal behavior. We compute how the asymptotic misadjustment, measuring the average asymptotic distance from a stable fixed point of the ideal behavior, scales as a function of the learning parameter and the number of training patterns. Knowing the asymptotic misadjustment, we calculate the typical number of learning steps necessary to generate a weight within order /spl epsiv/ of this fixed point, both with fixed and time-dependent learning parameters. We conclude that almost-cyclic learning (learning with random cycles) is a better alternative for batch-mode learning than cyclic learning (learning with a fixed cycle).Keywords
This publication has 15 references indexed in Scilit:
- On-Line Learning with Time-Correlated PatternsEurophysics Letters, 1994
- Stochastic dynamics of learning with momentum in neural networksJournal of Physics A: General Physics, 1994
- Computing second derivatives in feed-forward networks: a reviewIEEE Transactions on Neural Networks, 1994
- On stochastic dynamics of supervised learningJournal of Physics A: General Physics, 1993
- Cooling schedules for learning in neural networksPhysical Review E, 1993
- Stochastic dynamics of supervised learningJournal of Physics A: General Physics, 1993
- Learning in neural networks with local minimaPhysical Review A, 1992
- Optimization for training neural netsIEEE Transactions on Neural Networks, 1992
- On the Convergence of the LMS Algorithm with Adaptive Learning Rate for Linear Feedforward NetworksNeural Computation, 1991
- Self-organized formation of topologically correct feature mapsBiological Cybernetics, 1982