Cost-sensitive boosting algorithms: Do we really need them?

被引:41
|
作者
Nikolaou, Nikolaos [1 ]
Edakunni, Narayanan [1 ]
Kull, Meelis [2 ]
Flach, Peter [2 ]
Brown, Gavin [1 ]
机构
[1] Univ Manchester, Sch Comp Sci, Kilburn Bldg,Oxford Rd, Manchester M13 9PL, Lancs, England
[2] Univ Bristol, Dept Comp Sci, Merchant Venturers Bldg,Woodland Rd, Bristol BS8 1UB, Avon, England
基金
英国工程与自然科学研究理事会;
关键词
Boosting; Cost-sensitive; Class imbalance; Classifier calibration;
D O I
10.1007/s10994-016-5572-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We provide a unifying perspective for two decades of work on cost-sensitive Boosting algorithms. When analyzing the literature 1997-2016, we find 15 distinct cost-sensitive variants of the original algorithm; each of these has its own motivation and claims to superiority-so who should we believe? In this work we critique the Boosting literature using four theoretical frameworks: Bayesian decision theory, the functional gradient descent view, margin theory, and probabilistic modelling. Our finding is that only three algorithms are fully supported-and the probabilistic model view suggests that all require their outputs to be calibrated for best performance. Experiments on 18 datasets across 21 degrees of imbalance support the hypothesis-showing that once calibrated, they perform equivalently, and outperform all others. Our final recommendation-based on simplicity, flexibility and performance-is to use the original Adaboost algorithm with a shifted decision threshold and calibrated probability estimates.
引用
收藏
页码:359 / 384
页数:26
相关论文
共 50 条
  • [1] Cost-sensitive boosting algorithms: Do we really need them?
    Nikolaos Nikolaou
    Narayanan Edakunni
    Meelis Kull
    Peter Flach
    Gavin Brown
    Machine Learning, 2016, 104 : 359 - 384
  • [2] Cost-sensitive boosting algorithms as gradient descent
    Cai, Qu-Tang
    Song, Yang-Qui
    Zhang, Chang-Shui
    2008 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-12, 2008, : 2009 - 2012
  • [3] Parameter inference of cost-sensitive boosting algorithms
    Sun, YM
    Wong, AKC
    Wang, Y
    MACHINE LEARNING AND DATA MINING IN PATTERN RECOGNITION, PROCEEDINGS, 2005, 3587 : 21 - 30
  • [4] Cost-sensitive classification with cost uncertainty: do we need surrogate losses?
    Komisarenko, Viacheslav
    Kull, Meelis
    MACHINE LEARNING, 2025, 114 (06)
  • [5] Cost-Sensitive Boosting
    Masnadi-Shirazi, Hamed
    Vasconcelos, Nuno
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (02) : 294 - 309
  • [6] VITAMINS, DO WE REALLY NEED THEM
    BLOCK, MB
    ARIZONA MEDICINE, 1980, 37 (04) : 281 - 281
  • [7] Chaperones: Do we really need them?
    Lin, PC
    Mandel, L
    Soules, MR
    FERTILITY AND STERILITY, 2004, 82 : S139 - S140
  • [8] Practice guidelines: Do we really need them?
    Amerling, R
    PERITONEAL DIALYSIS INTERNATIONAL, 2005, 25 (02): : 140 - 141
  • [9] Boosting cost-sensitive trees
    Ting, KM
    Zheng, ZJ
    DISCOVERY SCIENCE, 1998, 1532 : 244 - 255
  • [10] Chromosome replication origins: Do we really need them?
    Michel, Benedicte
    Bernander, Rolf
    BIOESSAYS, 2014, 36 (06) : 585 - 590