Minimax nonparametric classification .I. Rates of convergence
- 1 November 1999
- journal article
- Published by Institute of Electrical and Electronics Engineers (IEEE) in IEEE Transactions on Information Theory
- Vol. 45 (7) , 2271-2284
- https://doi.org/10.1109/18.796368
Abstract
This paper studies minimax aspects of nonparametric classification. We first study minimax estimation of the conditional probability of a class label, given the feature variable. This function, say f, is assumed to be in a general nonparametric class. We show the minimax rate of convergence under square L/sub 2/ loss is determined by the massiveness of the class as measured by metric entropy. The second part of the paper studies minimax classification. The loss of interest is the difference between the probability of misclassification of a classifier and that of the Bayes decision. As is well known, an upper bound on risk for estimating f gives an upper bound on the risk for classification, but the rate is known to be suboptimal for the class of monotone functions. This suggests that one does not have to estimate f well in order to classify well. However, we show that the two problems are in fact of the same difficulty in terms of rates of convergence under a sufficient condition, which is satisfied by many function classes including Besov (Sobolev), Lipschitz, and bounded variation. This is somewhat surprising in view of a result of Devroye, Gorfi, and Lugosi (see A Probabilistic Theory of Pattern Recognition, New York: Springer-Verlag, 1996).Keywords
This publication has 28 references indexed in Scilit:
- Minimax nonparametric classification. II. Model selection for adaptationIEEE Transactions on Information Theory, 1999
- Information-theoretic determination of minimax rates of convergenceThe Annals of Statistics, 1999
- Consistency of data-driven histogram methods for density estimation and classificationThe Annals of Statistics, 1996
- The rates of convergence of kernel regression estimates and classification rulesIEEE Transactions on Information Theory, 1986
- Rates of Convergence of Minimum Distance Estimators and Kolmogorov's EntropyThe Annals of Statistics, 1985
- Asymptotic efficiency of classifying procedures using the Hermite series estimate of multivariate probability densities (Corresp.)IEEE Transactions on Information Theory, 1981
- The rate of convergence ofk_n-NN regression estimates and classification rules (Corresp.)IEEE Transactions on Information Theory, 1981
- Entropy numbers of embedding maps between Besov spaces with an application to eigenvalue problemsProceedings of the Royal Society of Edinburgh: Section A Mathematics, 1981
- Consistent Nonparametric RegressionThe Annals of Statistics, 1977
- Metric entropy and approximationBulletin of the American Mathematical Society, 1966