A New Class of Incremental Gradient Methods for Least Squares Problems
- 1 November 1997
- journal article
- Published by Society for Industrial & Applied Mathematics (SIAM) in SIAM Journal on Optimization
- Vol. 7 (4) , 913-926
- https://doi.org/10.1137/s1052623495287022
Abstract
The least mean squares (LMS) method for linear least squares problems differs from the steepest descent method in that it processes data blocks one-by-one, with intermediate adjustment of the parameter vector under optimization. This mode of operation often leads to faster convergence when far from the eventual limit and to slower (sublinear) convergence when close to the optimal solution. We embed both LMS and steepest descent, as well as other intermediate methods, within a one-parameter class of algorithms, and we propose a hybrid class of methods that combine the faster early convergence rate of LMS with the faster ultimate linear convergence rate of steepest descent. These methods are well suited for neural network training problems with large data sets. Furthermore, these methods allow the effective use of scaling based, for example, on diagonal or other approximations of the Hessian matrix.Keywords
This publication has 14 references indexed in Scilit:
- Incremental Least Squares Methods and the Extended Kalman FilterSIAM Journal on Optimization, 1996
- The Iterated Kalman Smoother as a Gauss–Newton MethodSIAM Journal on Optimization, 1994
- A class of unconstrained minimization methods for neural network trainingOptimization Methods and Software, 1994
- Convergence properties of backpropagation for neural nets via theory of stochastic gradient methods. Part 1Optimization Methods and Software, 1994
- Serial and parallel backpropagation convergence via nonmonotone perturbed minimizationOptimization Methods and Software, 1994
- Analysis of an approximate gradient projection method with applications to the backpropagation algorithmOptimization Methods and Software, 1994
- On the Convergence of the LMS Algorithm with Adaptive Learning Rate for Linear Feedforward NetworksNeural Computation, 1991
- Stochastic Approximation Methods for Constrained and Unconstrained SystemsPublished by Springer Nature ,1978
- Analysis of recursive stochastic algorithmsIEEE Transactions on Automatic Control, 1977
- New least-square algorithmsJournal of Optimization Theory and Applications, 1976