Learning in linear neural networks: The validity of the annealed approximation
- 1 August 1992
- journal article
- research article
- Published by American Physical Society (APS) in Physical Review A
- Vol. 46 (4) , 2124-2130
- https://doi.org/10.1103/physreva.46.2124
Abstract
A statistical-mechanics approach has been recently used to develop a theoretical framework to describe the process of supervised learning and the emergence of generalization ability in layered neural networks. This theory has yielded a powerful tool: a recursion relation that predicts the generalization ability of trained networks as a function of the size of the training set. The recursion relation results from an annealed approximation to the correct quenched averaging over all possible training sets of fixed size. Here we investigate a simple learning problem for which both the quenched and annealed calculations are performed analytically. Comparison between the corresponding results establishes the range of validity of the annealed approximation.Keywords
This publication has 9 references indexed in Scilit:
- GENERALIZATION ERROR AND DYNAMICAL EFFECTS IN A TWO-DIMENSIONAL PATCHES DETECTORInternational Journal of Neural Systems, 1991
- Learning in a Two-Layer Neural Network of Edge DetectorsEurophysics Letters, 1990
- Statistical mechanics of a multilayered neural networkPhysical Review Letters, 1990
- A statistical approach to learning and generalization in layered neural networksProceedings of the IEEE, 1990
- Learning from examples in large neural networksPhysical Review Letters, 1990
- Exhaustive LearningNeural Computation, 1990
- Learning from Examples in a Single-Layer Neural NetworkEurophysics Letters, 1990
- Parallel Distributed ProcessingPublished by MIT Press ,1986
- On the connection between spin glasses and gauge field theoriesPhysics Reports, 1980