On weak base hypotheses and their implications for boosting regression and classification

26Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

When studying the training error and the prediction error for boosting, it is often assumed that the hypotheses returned by the base learner are weakly accurate, or are able to beat a random guesser by a certain amount of difference. It has been an open question how much this difference can be, whether it will eventually disappear in the boosting process or be bounded by a positive amount. This question is crucial for the behavior of both the training error and the prediction error. In this paper we study this problem and show affirmatively that the amount of improvement over the random guesser will be at least a positive amount for almost all possible sample realizations and for most of the commonly used base hypotheses. This has a number of implications for the prediction error, including, for example, that boosting forever may not be good and regularization may be necessary. The problem is studied by first considering an analog of AdaBoost in regression, where we study similar properties and find that, for good performance, one cannot hope to avoid regularization by just adopting the boosting device to regression.

References Powered by Scopus

Bagging predictors

19182Citations
N/AReaders
Get full text

A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting

13319Citations
N/AReaders
Get full text

Matching Pursuits With Time-Frequency Dictionaries

8403Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Gradient boosting machines, a tutorial

2347Citations
N/AReaders
Get full text

Process consistency for AdaBoost

87Citations
N/AReaders
Get full text

Bandwidth choice for nonparametric classification

51Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Jiang, W. (2002). On weak base hypotheses and their implications for boosting regression and classification. Annals of Statistics, 30(1), 51–73. https://doi.org/10.1214/aos/1015362184

Readers over time

‘11‘14‘15‘17‘18‘19‘23‘2400.751.52.253

Readers' Seniority

Tooltip

Researcher 5

50%

PhD / Post grad / Masters / Doc 3

30%

Professor / Associate Prof. 2

20%

Readers' Discipline

Tooltip

Computer Science 5

63%

Neuroscience 1

13%

Physics and Astronomy 1

13%

Engineering 1

13%

Save time finding and organizing research with Mendeley

Sign up for free
0