Supervised projection approach for boosting classifiers

被引:19
作者
Garcia-Pedrajas, Nicolas [1 ]
机构
[1] Univ Cordoba, Dept Comp & Numer Anal, E-14071 Cordoba, Spain
关键词
Classification; Ensembles of classifiers; Boosting; Supervised projections; FEATURE-EXTRACTION; NETWORK; CLASSIFICATION; MARGIN; ALGORITHMS; ENSEMBLES; PURSUIT;
D O I
10.1016/j.patcog.2008.12.023
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper we present a new approach for boosting methods for the construction of ensembles of classifiers. The approach is based on using the distribution given by the weighting scheme of boosting to construct a non-linear supervised projection of the original variables, instead of using the weights of the instances to train the next classifier. With this method we construct ensembles that are able to achieve a better generalization error and are more robust to noise presence. It has been proved that AdaBoost method is able to improve the margin of the instances achieved by the ensemble. Moreover, its practical success has been partially explained by this margin maximization property. However, in noisy problems, likely to occur in real-world applications, the maximization of the margin of wrong instances or outliers can lead to poor generalization. We propose an alternative approach. where the distribution of the weights given by the boosting algorithm is used to get a supervised projection. Then, the supervised projection is used to train the next classifier using a uniform distribution of the training instances. The proposed approach is compared with three boosting techniques, namely AdaBoost, GentleBoost and MadaBoost, showing an improved performance on a large set of 55 problems from the LICI Machine Learning Repository, and less sensitiveness to noise in the class labels. The behavior of the proposed algorithm in terms of margin distribution and bias-variance decomposition is also studied. (C) 2009 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1742 / 1760
页数:19
相关论文
共 59 条
[1]  
[Anonymous], 1969, Statistical computation, DOI [10.1016/B978-0-12-498150-8.50024-0, DOI 10.1016/B978-0-12-498150-8.50024-0]
[2]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[3]  
Breiman L, 1996, MACH LEARN, V24, P49
[4]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[5]   Prediction games and arcing algorithms [J].
Breiman, L .
NEURAL COMPUTATION, 1999, 11 (07) :1493-1517
[6]  
Breiman L., 1996, 460 U CAL DEP STAT
[7]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[8]  
Cristianini N., 2000, An Introduction to Support Vector Machines and Other Kernel-based Learning Methods, DOI DOI 10.1017/CB09780511801389
[9]  
Demsar J, 2006, J MACH LEARN RES, V7, P1
[10]   An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization [J].
Dietterich, TG .
MACHINE LEARNING, 2000, 40 (02) :139-157