Improving Model Selection by Employing the Test Data

被引:0
|
作者
Westphal, Max [1 ]
Brannath, Werner [1 ]
机构
[1] Univ Bremen, Fac Math & Comp Sci 3, Inst Stat, Bremen, Germany
来源
INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97 | 2019年 / 97卷
关键词
MULTIPLE COMPARISONS; OVER-OPTIMISM; BIOINFORMATICS; INTELLIGENCE; INFERENCE; DESIGN;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Model selection and evaluation are usually strictly separated by means of data splitting to enable an unbiased estimation and a simple statistical inference for the unknown generalization performance of the final prediction model. We investigate the properties of novel evaluation strategies, namely when the final model is selected based on empirical performances on the test data. To guard against selection induced overoptimism, we employ a parametric multiple test correction based on the approximate multivariate distribution of performance estimates. Our numerical experiments involve training common machine learning algorithms (EN, CART, SVM, XGB) on various artificial classification tasks. At its core, our proposed approach improves model selection in terms of the expected final model performance without introducing overoptimism. We furthermore observed a higher probability for a successful evaluation study, making it easier in practice to empirically demonstrate a sufficiently high predictive performance.
引用
收藏
页数:10
相关论文
共 50 条