Advances in feedforward neural networks: demystifying knowledge acquiring black boxes
- 1 April 1996
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE Transactions on Knowledge and Data Engineering
- Vol. 8 (2) , 211-226
- https://doi.org/10.1109/69.494162
Abstract
We survey research of recent years on the supervised training of feedforward neural networks. The goal is to expose how the networks work, how to engineer them so they can learn data with less extraneous noise, how to train them efficiently, and how to assure that the training is valid. The scope covers gradient descent and polynomial line search, from backpropagation through conjugate gradients and quasi-Newton methods. There is a consensus among researchers that adaptive step gains (learning rates) can stabilize and accelerate convergence and that a good starting weight set improves both the training speed and the learning quality. The training problem includes both the design of a network function and the fitting of the function to a set of input and output data points by computing a set of coefficient weights. The form of the function can be adjusted by adjoining new neurons and pruning existing ones and setting other parameters such as biases and exponential rates. Our exposition reveals several useful results that are readily implementable.Keywords
This publication has 42 references indexed in Scilit:
- A fast algorithm for the hidden units in a multilayer perceptronPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2005
- Some n-bit parity problems are solvable by feedforward networks with less than n hidden unitsPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2005
- Removal of hidden units and weights for back propagation networksPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2005
- Generalization by weight-elimination applied to currency exchange rate predictionPublished by Institute of Electrical and Electronics Engineers (IEEE) ,2002
- BACKPROPAGATION LEARNING FOR MULTILAYER FEED-FORWARD NEURAL NETWORKS USING THE CONJUGATE GRADIENT METHODInternational Journal of Neural Systems, 1991
- PREDICTING THE FUTURE: A CONNECTIONIST APPROACHInternational Journal of Neural Systems, 1990
- Connectionist learning proceduresArtificial Intelligence, 1989
- Increased rates of convergence through learning rate adaptationNeural Networks, 1988
- Neural net pruning-why and howPublished by Institute of Electrical and Electronics Engineers (IEEE) ,1988
- Methods of conjugate gradients for solving linear systemsJournal of Research of the National Bureau of Standards, 1952