Process consistency for AdaBoost
Open Access
- 1 February 2004
- journal article
- Published by Institute of Mathematical Statistics in The Annals of Statistics
- Vol. 32 (1) , 13-29
- https://doi.org/10.1214/aos/1079120128
Abstract
Recent experiments and theoretical studies show that AdaBoost can overfit in the limit of large time. If running the algorithm forever is suboptimal, a natural question is how low can the prediction error be during the process of AdaBoost? We show under general regularity conditions that during the process of AdaBoost a consistent prediction is generated, which has the prediction error approximating the optimal Bayes error as the sample size increases. This result suggests that, while running the algorithm forever can be suboptimal, it is reasonable to expect that some regularization method via truncation of the process may lead to a near-optimal performance for sufficiently large sample size.Keywords
This publication has 7 references indexed in Scilit:
- On weak base Hypotheses and their implications for boosting regression and classificationThe Annals of Statistics, 2002
- Additive logistic regression: a statistical view of boosting (With discussion and a rejoinder by the authors)The Annals of Statistics, 2000
- Theoretical Views of BoostingPublished by Springer Nature ,1999
- Boosting the margin: a new explanation for the effectiveness of voting methodsThe Annals of Statistics, 1998
- Arcing classifier (with discussion and a rejoinder by the author)The Annals of Statistics, 1998
- A Decision-Theoretic Generalization of On-Line Learning and an Application to BoostingJournal of Computer and System Sciences, 1997
- A Probabilistic Theory of Pattern RecognitionPublished by Springer Nature ,1996