The general inefficiency of batch training for gradient descent learning
Top Cited Papers
- 1 December 2003
- journal article
- research article
- Published by Elsevier in Neural Networks
- Vol. 16 (10) , 1429-1451
- https://doi.org/10.1016/s0893-6080(03)00138-2
Abstract
No abstract availableKeywords
This publication has 7 references indexed in Scilit:
- The need for small learning rates on large problemsPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2002
- New results on recurrent network training: unifying the algorithms and accelerating convergenceIEEE Transactions on Neural Networks, 2000
- A scaled conjugate gradient algorithm for fast supervised learningNeural Networks, 1993
- Speaker-independent isolated digit recognition: Multilayer perceptrons vs. Dynamic time warpingNeural Networks, 1990
- Self-organized formation of topologically correct feature mapsBiological Cybernetics, 1982
- Stochastic Estimation of the Maximum of a Regression FunctionThe Annals of Mathematical Statistics, 1952
- A Stochastic Approximation MethodThe Annals of Mathematical Statistics, 1951