Improving Neural Network Generalization on Data-limited Regression with Doubly-Robust Boosting

被引:0
|
作者
Wang, Hao [1 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Enhancing the generalization performance of neural networks given limited data availability remains a formidable challenge, due to the model selection trade-off between training error and generalization gap. To handle this challenge, we present a posterior optimization issue, specifically designed to reduce the generalization error of trained neural networks. To operationalize this concept, we propose a Doubly-Robust Boosting machine (DRBoost) which consists of a statistical learner and a zero-order optimizer. The statistical learner reduces the model capacity and thus the generalization gap; the zero-order optimizer minimizes the training error in a gradient-free manner. The two components cooperate to reduce the generalization error of a fully trained neural network in a doubly robust manner. Furthermore, the statistical learner alleviates the multi-collinearity in the discriminative layer and enhances the generalization performance. The zero-order optimizer eliminates the reliance on gradient calculation and offers more flexibility in learning objective selection. Experiments demonstrate that DRBoost improves the generalization performance of various prevalent neural network backbones effectively.
引用
收藏
页码:20821 / 20829
页数:9
相关论文
共 50 条
  • [31] Robust neural network regression for offline and online learning
    Briegel, T
    Tresp, V
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 12, 2000, 12 : 407 - 413
  • [32] Improving data-limited stock assessment with sporadic stock index information in stock reduction analysis
    Guan, Lisha
    Chen, Yong
    Boenish, Robert
    Jin, Xianshi
    Shan, Xiujuan
    CANADIAN JOURNAL OF FISHERIES AND AQUATIC SCIENCES, 2020, 77 (05) : 857 - 868
  • [33] Knowledge Transferred Fine-Tuning: Convolutional Neural Network Is Born Again With Anti-Aliasing Even in Data-Limited Situations
    Suzuki, Satoshi
    Takeda, Shoichiro
    Makishima, Naoki
    Ando, Atsushi
    Masumura, Ryo
    Shouno, Hayaru
    IEEE ACCESS, 2022, 10 : 68384 - 68396
  • [34] Adversarially Reprogramming Pretrained Neural Networks for Data-limited and Cost-efficient Malware Detection
    Chen, Lingwei
    Li, Xiaoting
    Wu, Dinghao
    PROCEEDINGS OF THE 2022 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2022, : 693 - 701
  • [35] Neural Network Classification for Improving Continuous Regression Testing
    Marijan, Dusica
    Gotlieb, Arnaud
    Sapkota, Abhijeet
    2020 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE TESTING (AITEST), 2020, : 123 - 124
  • [36] A modified boosting based neural network ensemble method for regression and forecasting
    Wang, Li
    Zhu, Xuefeng
    ICIEA 2007: 2ND IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS, VOLS 1-4, PROCEEDINGS, 2007, : 1280 - 1285
  • [37] Improving the generalization performance of RBF neural networks using a linear regression technique
    Lin, C. L.
    Wang, J. F.
    Chen, C. Y.
    Chen, C. W.
    Yen, C. W.
    EXPERT SYSTEMS WITH APPLICATIONS, 2009, 36 (10) : 12049 - 12053
  • [38] Spatial Data Augmentation: Improving the Generalization of Neural Networks for Pansharpening
    Chen, Lihui
    Vivone, Gemine
    Nie, Zihao
    Chanussot, Jocelyn
    Yang, Xiaomin
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [39] On the training error and generalization error of neural network regression without identifiability
    Hagiwara, K
    KNOWLEDGE-BASED INTELLIGENT INFORMATION ENGINEERING SYSTEMS & ALLIED TECHNOLOGIES, PTS 1 AND 2, 2001, 69 : 1575 - 1579
  • [40] Using the hermite regression algorithm to improve the generalization capability of a neural network
    Pilato, G
    Sorbello, F
    Vassallo, G
    NEURAL NETS - WIRN VIETRI-99, 1999, : 296 - 301