Two-layer perceptrons at saturation
- 1 June 1992
- journal article
- research article
- Published by American Physical Society (APS) in Physical Review A
- Vol. 45 (12) , 8924-8937
- https://doi.org/10.1103/physreva.45.8924
Abstract
We study multilayer networks which implement a fixed Boolean function from the hidden layer to the output, and have a fully connected architecture from the input to the hidden layer. We analyze the organization of the first layer of weights and the capacity of the network by a statistical mechanical approach. The mean-field equations that govern the behavior of a network at saturation are derived assuming a replica symmetric solution. The theory is applied for a detailed analysis of the learning ability of an AND machine. We calculate the maximal capacity, and the overlap between the subnetworks. Attention is paid to the organization of internal representations at saturation. The results are compared with detailed numerical simulations, and with bounds on the capacity. Good agreement is found. The behavior of the system depends on the ratio between the number of patterns with a (+) and (-) output. We compare the fully connected architecture with networks that have nonoverlapping receptive fields from the input to the hidden layer. The mean-field equations for this architecture are also presented.Keywords
This publication has 12 references indexed in Scilit:
- Broken symmetries in multilayered perceptronsPhysical Review A, 1992
- Information Theory of a Multilayer Neural Network with Discrete WeightsEurophysics Letters, 1992
- On the capacity of neural networks with binary weightsJournal of Physics A: General Physics, 1992
- Storage Capacity of a Multilayer Neural Network with Binary WeightsEurophysics Letters, 1991
- Statistical mechanics of a multilayered neural networkPhysical Review Letters, 1990
- Optimal basins of attraction in randomly sparse neural network modelsJournal of Physics A: General Physics, 1989
- Three unfinished works on the optimal storage capacity of networksJournal of Physics A: General Physics, 1989
- Bounds on the learning capacity of some multi-layer networksBiological Cybernetics, 1989
- The space of interactions in neural network modelsJournal of Physics A: General Physics, 1988
- Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern RecognitionIEEE Transactions on Electronic Computers, 1965