On Oblique Random Forests

被引:0
|
作者
Menze, Bjoern H. [1 ]
Kelm, B. Michael [1 ]
Splitthoff, Daniel N. [1 ]
Koethe, Ullrich [1 ]
Hamprecht, Fred A. [1 ]
机构
[1] Heidelberg Univ, Interdisciplinary Ctr Sci Comp, Heidelberg, Germany
关键词
FEATURE-SELECTION; DECISION TREES; CLASSIFICATION; ENSEMBLES;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In his original paper on random forests, Breiman proposed two different decision tree ensembles: one generated from "orthogonal" trees with thresholds on individual features in every split, and one from "oblique" trees separating the feature space by randomly oriented hyperplanes. In spite of a rising interest in the random forest framework, however, ensembles built from orthogonal trees (RF) have gained most, if not all, attention so far. In the present work we propose to employ "oblique" random forests (oRF) built from multivariate trees which explicitly learn optimal split directions at internal nodes using linear discriminative models, rather than using random coefficients as the original oRF. This oRF outperforms RF, as well as other classifiers, on nearly all data sets but those with discrete factorial features. Learned node models perform distinctively better than random splits. An oRF feature importance score shows to be preferable over standard RF feature importance scores such as Gini or permutation importance. The topology of the oRF decision space appears to be smoother and better adapted to the data, resulting in improved generalization performance. Overall, the oRF propose here may be preferred over standard RF on most learning tasks involving numerical and spectral data.
引用
收藏
页码:453 / 469
页数:17
相关论文
共 50 条
  • [41] Streaming random forests
    Abdulsalam, Hanady
    Skillicorn, David B.
    Martin, Patrick
    IDEAS 2007: 11TH INTERNATIONAL DATABASE ENGINEERING AND APPLICATIONS SYMPOSIUM, PROCEEDINGS, 2007, : 225 - 232
  • [42] Random Decomposition Forests
    Chien, Chun-Han
    Chen, Hwann-Tzong
    2013 SECOND IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR 2013), 2013, : 687 - 691
  • [43] Uplift Random Forests
    Guelman, Leo
    Guillen, Montserrat
    Perez-Marin, Ana M.
    CYBERNETICS AND SYSTEMS, 2015, 46 (3-4) : 230 - 248
  • [44] Subtractive random forests
    Broutin, Nicolas
    Devroye, Luc
    Lugosi, Gabor
    Oliveira, Roberto Imbuzeiro
    ALEA-LATIN AMERICAN JOURNAL OF PROBABILITY AND MATHEMATICAL STATISTICS, 2024, 21 : 575 - 591
  • [45] Neural Random Forests
    Gérard Biau
    Erwan Scornet
    Johannes Welbl
    Sankhya A, 2019, 81 : 347 - 386
  • [46] Random subsequence forests
    He Z.
    Wang J.
    Jiang M.
    Hu L.
    Zou Q.
    Information Sciences, 2024, 667
  • [47] Compressing Random Forests
    Painsky, Amichai
    Rosset, Saharon
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2016, : 1131 - 1136
  • [48] Random forests for microarrays
    Cutler, Adele
    Stevens, John R.
    DNA MICROARRAYS, PART B: DATABASES AND STATISTICS, 2006, 411 : 422 - +
  • [49] On the profile of random forests
    Gittenberger, B
    MATHEMATICS AND COMPUTER SCIENCE II: ALGORITHMS, TREES, COMBINATORICS AND PROBABILITIES, 2002, : 279 - 293
  • [50] Random Forests with R
    Maindonald, John H.
    INTERNATIONAL STATISTICAL REVIEW, 2021,