Generalizing the PAC model: sample size bounds from metric dimension-based uniform convergence results
- 1 January 1989
- conference paper
- Published by Institute of Electrical and Electronics Engineers (IEEE)
Abstract
The probably approximately correct (PAC) model of learning from examples is generalized. The problem of learning functions from a set X into a set Y is considered, assuming only that the examples are generated by independent draws according to an unknown probability measure on X*Y. The learner's goal is to find a function in a given hypothesis space of functions from X into Y that on average give Y values that are close to those observed in random examples. The discrepancy is measured by a bounded real-valued loss function. The average loss is called the error of the hypothesis. A theorem on the uniform convergence of empirical error estimates to true error rates is given for certain hypothesis spaces, and it is shown how this implies learnability. A generalized notion of VC dimension that applies to classes of real-valued functions and a notion of capacity for classes of functions that map into a bounded metric space are given. These measures are used to bound the rate of convergence of empirical error estimates to true error rates, giving bounds on the sample size needed for learning using hypotheses in these classes. As an application, a distribution-independent uniform convergence result for certain classes of functions computed by feedforward neural nets is obtained. Distribution-specific uniform convergence results for classes of functions that are uniformly continuous on average are also obtained.<>Keywords
This publication has 15 references indexed in Scilit:
- Learnability and the Vapnik-Chervonenkis dimensionJournal of the ACM, 1989
- Automatic pattern recognition: a study of the probability of errorPublished by Institute of Electrical and Electronics Engineers (IEEE) ,1988
- Queries and concept learningMachine Learning, 1988
- ɛ-nets and simplex range queriesDiscrete & Computational Geometry, 1987
- A theory of the learnableCommunications of the ACM, 1984
- Convergence of Stochastic ProcessesPublished by Springer Nature ,1984
- The dimension of chaotic attractorsPhysica D: Nonlinear Phenomena, 1983
- Information Dimension and the Probabilistic Structure of ChaosZeitschrift für Naturforschung A, 1982
- Central Limit Theorems for Empirical MeasuresThe Annals of Probability, 1978
- 𝜀-entropy and 𝜀-capacity of sets in functional spacesPublished by American Mathematical Society (AMS) ,1961