Selecting concise training sets from clean data
- 1 March 1993
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE Transactions on Neural Networks
- Vol. 4 (2) , 305-318
- https://doi.org/10.1109/72.207618
Abstract
The authors derive a method for selecting exemplars for training a multilayer feedforward network architecture to estimate an unknown (deterministic) mapping from clean data, i.e., data measured either without error or with negligible error. The objective is to minimize the data requirement of learning. The authors choose a criterion for selecting training examples that works well in conjunction with the criterion used for learning, here, least squares. They proceed sequentially, selecting an example that, when added to the previous set of training examples and learned, maximizes the decrement of network squared error over the input space. When dealing with clean data and deterministic relationships, concise training sets that minimize the integrated squared bias (ISB) are desired. The ISB is used to derive a selection criterion for evaluating individual training examples, the DISB, that is maximized to select new exemplars. They conclude with graphical illustrations of the method, and demonstrate its use during network training. Experimental results indicate that training upon exemplars selected in this fashion can save computation in general purpose use as well.Keywords
This publication has 26 references indexed in Scilit:
- The Evidence Framework Applied to Classification NetworksNeural Computation, 1992
- Information-Based Objective Functions for Active Data SelectionNeural Computation, 1992
- A Resource-Allocating Network for Function InterpolationNeural Computation, 1991
- A Local Approach to Optimal QueriesPublished by Elsevier ,1991
- Response Surface Methodology: 1966-1988Technometrics, 1989
- Nonlinear RegressionWiley Series in Probability and Statistics, 1989
- The sequential generation of multiresponse d-optimal designs when the variance-covariance matrix is not knownCommunications in Statistics - Simulation and Computation, 1987
- A theory of the learnableCommunications of the ACM, 1984
- Optimal designs for nonparametric kernel regressionStatistics & Probability Letters, 1984
- Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern RecognitionIEEE Transactions on Electronic Computers, 1965