Randomizing outputs to increase prediction accuracy

被引:187
作者
Breiman, L [1 ]
机构
[1] Univ Calif Berkeley, Dept Stat, Berkeley, CA 94720 USA
关键词
ensemble; randomization; output variability;
D O I
10.1023/A:1007682208299
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bagging and boosting reduce error by changing both the inputs and outputs to form perturbed training sets, growing predictors on these perturbed training sets and combining them. An interesting question is whether it is possible to get comparable performance by perturbing the outputs alone. Two methods of randomizing outputs are experimented with. One is called output smearing and the other output flipping. Both are shown to consistently do better than bagging.
引用
收藏
页码:229 / 242
页数:14
相关论文
共 14 条
[1]  
AN G, 1996, NEURAL COMPUT, V6, P643
[2]  
Breiman L, 1996, ANN STAT, V24, P2350
[3]  
Breiman L, 1998, ANN STAT, V26, P801
[4]   Bagging predictors [J].
Breiman, L .
MACHINE LEARNING, 1996, 24 (02) :123-140
[5]  
BREIMAN L, 1997, 504 U CAL BERK STAT
[6]  
Breiman L., 1984, BIOMETRICS, DOI DOI 10.2307/2530946
[7]  
BREIMAN L, 1998, 534 U CAL BERK STAT
[8]  
DIETTERICH T, 1998, MACH LEARN, P1
[9]  
Freund Y., 1996, Machine Learning. Proceedings of the Thirteenth International Conference (ICML '96), P148
[10]   A decision-theoretic generalization of on-line learning and an application to boosting [J].
Freund, Y ;
Schapire, RE .
JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1997, 55 (01) :119-139