Computational capabilities of restricted two-layered perceptrons
- 1 July 1994
- journal article
- research article
- Published by American Physical Society (APS) in Physical Review E
- Vol. 50 (1) , 577-595
- https://doi.org/10.1103/physreve.50.577
Abstract
We study the extent to which fixing the second-layer weights reduces the capacity and generalization ability of a two-layer perceptron. Architectures with N inputs, K hidden units, and a single output are considered, with both overlapping and nonoverlapping receptive fields. We obtain from simulations one measure of the strength of a network—its critical capacity, . Using the ansatz ∝(-α to describe the manner in which the median learning time diverges as is approached, we estimate in a manner that does not depend on arbitrary impatience parameters. The c h i r learning algorithm is used in our simulations. For K=3 and overlapping receptive fields we show that the general machine is equivalent to the committee machine with the same architecture. For K=5 and the same connectivity the general machine is the union of four distinct networks with fixed second layer weights, of which the committee machine is the one with the highest . Since the capacity of the union of a finite set of machines equals that of the strongest constituent, the capacity of the general machine with K=5 equals that of the committee machine. We were not able to prove this for general K , but believe that it does hold. We investigated the internal representations used by different machines, and found that high correlations between the hidden units and the output reduce the capacity. Finally we studied the Boolean functions that can be realized by networks with fixed second layer weights. We discovered that two different machines implement two completely distinct sets of Boolean functions.
Keywords
This publication has 18 references indexed in Scilit:
- Numerical Study of Back-Propagation Learning Algorithms for Multilayer NetworksEurophysics Letters, 1993
- Two-layer perceptrons at saturationPhysical Review A, 1992
- Broken symmetries in multilayered perceptronsPhysical Review A, 1992
- Storage Capacity of a Multilayer Neural Network with Binary WeightsEurophysics Letters, 1991
- Statistical mechanics of a multilayered neural networkPhysical Review Letters, 1990
- Bounds on the learning capacity of some multi-layer networksBiological Cybernetics, 1989
- Neural nets for adaptive filtering and adaptive pattern recognitionComputer, 1988
- Optimal storage properties of neural network modelsJournal of Physics A: General Physics, 1988
- The space of interactions in neural network modelsJournal of Physics A: General Physics, 1988
- Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern RecognitionIEEE Transactions on Electronic Computers, 1965