On weak base hypotheses and their implications for boosting regression and classification

被引:24
|
作者
Jiang, WX [1 ]
机构
[1] Northwestern Univ, Dept Stat, Evanston, IL 60208 USA
来源
ANNALS OF STATISTICS | 2002年 / 30卷 / 01期
关键词
angular span; boosting; classification; error bounds; least squares regression; matching pursuit; nearest neighbor rule; overfit; prediction error; regularization; training error; weak hypotheses;
D O I
10.1214/aos/1015362184
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
When studying the training error and the prediction error for boosting, it is often assumed that the hypotheses returned by the base learner are weakly accurate, or are able to beat a random guesser by a certain amount of difference. It has been an open question how much this difference can be. whether it will eventually disappear in the boosting process or be bounded by a positive amount. This question is crucial for the behavior of both the training error and the prediction error. In this paper we study this problem and show affirmatively that the amount of improvement over the random guesser will be at least a positive amount for almost all possible sample realizations and for most of the commonly used base hypotheses. This has a number of implications for the prediction error, including, for example, that boosting forever may not be good and regularization may be necessary. The problem is studied by first considering an analog of AdaBoost in regression, where we study similar properties and find that, for good performance, one cannot hope to avoid regularization by just adopting the boosting device to regression.
引用
收藏
页码:51 / 73
页数:23
相关论文
共 50 条
  • [1] Some results on weakly accurate base learners for boosting regression and classification
    Jiang, WX
    MULTIPLE CLASSIFIER SYSTEMS, 2000, 1857 : 87 - 96
  • [2] Weak hypotheses and boosting for generic object detection and recognition
    Opelt, A
    Fussenegger, M
    Pinz, A
    Auer, P
    COMPUTER VISION - ECCV 2004, PT 2, 2004, 3022 : 71 - 84
  • [3] Gradient and Newton boosting for classification and regression
    Sigrist, Fabio
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 167
  • [4] Gradient and Newton boosting for classification and regression
    Sigrist, Fabio
    Expert Systems with Applications, 2021, 167
  • [5] Boosting weak classifiers for visual tracking based on kernel regression
    Ma, Bo
    Ma, Weizhang
    MIPPR 2011: AUTOMATIC TARGET RECOGNITION AND IMAGE ANALYSIS, 2011, 8003
  • [6] Multi-resolution boosting for classification and regression problems
    Chandan K. Reddy
    Jin-Hyeong Park
    Knowledge and Information Systems, 2011, 29 : 435 - 456
  • [7] Multi-resolution Boosting for Classification and Regression Problems
    Reddy, Chandan K.
    Park, Jin-Hyeong
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PROCEEDINGS, 2009, 5476 : 196 - +
  • [8] QBoost: Predicting quantiles with boosting for regression and binary classification
    Zheng, Songfeng
    EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (02) : 1687 - 1697
  • [9] Local Boosting of Decision Stumps for Regression and Classification Problems
    Kotsiantis, S. B.
    Kanellopoulos, D.
    Pintelas, P. E.
    JOURNAL OF COMPUTERS, 2006, 1 (04) : 30 - 37
  • [10] Multi-resolution boosting for classification and regression problems
    Reddy, Chandan K.
    Park, Jin-Hyeong
    KNOWLEDGE AND INFORMATION SYSTEMS, 2011, 29 (02) : 435 - 456