Neural Network Studies. 2. Variable Selection
- 1 January 1996
- journal article
- research article
- Published by American Chemical Society (ACS) in Journal of Chemical Information and Computer Sciences
- Vol. 36 (4) , 794-803
- https://doi.org/10.1021/ci950204c
Abstract
Quantitative structure−activity relationship (QSAR) studies usually require an estimation of the relevance of a very large set of initial variables. Determination of the most important variables allows theoretically a better generalization by all pattern recognition methods. This study introduces and investigates five pruning algorithms designed to estimate the importance of input variables in feed-forward artificial neural network trained by back propagation algorithm (ANN) applications and to prune nonrelevant ones in a statistically reliable way. The analyzed algorithms performed similar variable estimations for simulated data sets, but differences were detected for real QSAR examples. Improvement of ANN prediction ability was shown after the pruning of redundant input variables. The statistical coefficients computed by ANNs for QSAR examples were better than those of multiple linear regression. Restrictions of the proposed algorithms and the potential use of ANNs are discussed.Keywords
This publication has 20 references indexed in Scilit:
- Neural network studies. 1. Comparison of overfitting and overtrainingJournal of Chemical Information and Computer Sciences, 1995
- Pruning from Adaptive RegularizationNeural Computation, 1994
- HIV-1 Reverse Transcriptase Inhibitor Design Using Artificial Neural NetworksJournal of Medicinal Chemistry, 1994
- Application of neural networks to a small dataset structure-activity relationshipJournal of Molecular Graphics, 1994
- Limitations of Functional‐Link Nets as Applied to QSAR Data AnalysisQuantitative Structure-Activity Relationships, 1994
- Statistics using neural networks: chance effectsJournal of Medicinal Chemistry, 1993
- Pruning algorithms-a surveyIEEE Transactions on Neural Networks, 1993
- A novel method to analyse response patterns of taste neurons by artificial neural networksNeuroReport, 1992
- Benefits of gain: speeded learning and minimal hidden layers in back-propagation networksIEEE Transactions on Systems, Man, and Cybernetics, 1991
- Using Relevance to Reduce Network Size AutomaticallyConnection Science, 1989