Ensembles of Overfit and Overconfident Forecasts

被引:49
|
作者
Grushka-Cockayne, Yael [1 ]
Jose, Victor Richmond R. [2 ]
Lichtendahl, Kenneth C., Jr. [1 ]
机构
[1] Univ Virginia, Darden Sch Business, Charlottesville, VA 22903 USA
[2] Georgetown Univ, McDonough Sch Business, Washington, DC 20057 USA
关键词
wisdom of crowds; base-rate neglect; linear opinion pool; trimmed opinion pool; hit rate; calibration; random forest; data science; PROBABILITY; AGGREGATION; CALIBRATION; ACCURACY; OPINIONS; WISDOM;
D O I
10.1287/mnsc.2015.2389
中图分类号
C93 [管理学];
学科分类号
12 ; 1201 ; 1202 ; 120202 ;
摘要
Firms today average forecasts collected from multiple experts and models. Because of cognitive biases, strategic incentives, or the structure of machine-learning algorithms, these forecasts are often overfit to sample data and are overconfident. Little is known about the challenges associated with aggregating such forecasts. We introduce a theoretical model to examine the combined effect of overfitting and overconfidence on the average forecast. Their combined effect is that the mean and median probability forecasts are poorly calibrated with hit rates of their prediction intervals too high and too low, respectively. Consequently, we prescribe the use of a trimmed average, or trimmed opinion pool, to achieve better calibration. We identify the random forest, a leading machine-learning algorithm that pools hundreds of overfit and overconfident regression trees, as an ideal environment for trimming probabilities. Using several known data sets, we demonstrate that trimmed ensembles can significantly improve the random forest's predictive accuracy.
引用
收藏
页码:1110 / 1130
页数:21
相关论文
共 50 条
  • [1] Ensembles of nearest neighbor forecasts
    Yankov, Dragomir
    DeCoste, Dennis
    Keogh, Eamonn
    MACHINE LEARNING: ECML 2006, PROCEEDINGS, 2006, 4212 : 545 - 556
  • [2] United We Stand: Using Epoch-Wise Agreement of Ensembles to Combat Overfit
    Stern, Uri
    Shwartz, Daniel
    Weinshall, Daphna
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 15075 - 15082
  • [3] Visualizing Forecasts of Neural Network Ensembles
    von Metthenheim, Hans-Joerg
    Koepp, Cornelius
    Breitner, Michael H.
    OPERATIONS RESEARCH PROCEEDINGS 2011, 2012, : 573 - 578
  • [4] Improving SCIPUFF Dispersion Forecasts with NWP Ensembles
    Lee, Jared A.
    Peltier, L. Joel
    Haupt, Sue Ellen
    Wyngaard, John C.
    Stauffer, David R.
    Deng, Aijun
    JOURNAL OF APPLIED METEOROLOGY AND CLIMATOLOGY, 2009, 48 (11) : 2305 - 2319
  • [5] To Overfit, or Not to Overfit: Improving the Performance of Deep Learning-Based SCA
    Rezaeezade, Azade
    Perin, Guilherme
    Picek, Stjepan
    PROGRESS IN CRYPTOLOGY - AFRICACRYPT 2022, 2022, 13503 : 397 - 421
  • [6] Improving Precipitation Forecasts by Generating Ensembles through Postprocessing
    Shrestha, Durga Lal
    Robertson, David E.
    Bennett, James C.
    Wang, Q. J.
    MONTHLY WEATHER REVIEW, 2015, 143 (09) : 3642 - 3663
  • [7] The use of ensembles to identify forecasts with small and large uncertainty
    Toth, Z
    Zhu, YJ
    Marchok, T
    WEATHER AND FORECASTING, 2001, 16 (04) : 463 - 477
  • [8] Optimizing Analog Ensembles for Sub-Daily Precipitation Forecasts
    Jeworrek, Julia
    West, Gregory
    Stull, Roland
    ATMOSPHERE, 2022, 13 (10)
  • [9] Improving week-2 forecasts with multimodel reforecast ensembles
    Whitaker, Jeffrey S.
    Wei, Xue
    Vitart, Frederic
    MONTHLY WEATHER REVIEW, 2006, 134 (08) : 2279 - 2284
  • [10] Hybrid predictive ensembles: Synergies between human and computational forecasts
    Hong L.
    Lamberson P.J.
    Page S.E.
    Journal of Social Computing, 2021, 2 (02): : 89 - 102