Regularizing Deep Neural Networks by Noise: Its Interpretation and Optimization

被引:0
|
作者
Noh, Hyeonwoo [1 ]
You, Tackgeun [1 ]
Mun, Jonghwan [1 ]
Han, Bohyung [1 ]
机构
[1] POSTECH, Dept Comp Sci & Engn, Pohang, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Overfitting is one of the most critical challenges in deep neural networks, and there are various types of regularization methods to improve generalization performance. Injecting noises to hidden units during training, e.g., dropout, is known as a successful regularizer, but it is still not clear enough why such training techniques work well in practice and how we can maximize their benefit in the presence of two conflicting objectives-optimizing to true data distribution and preventing overfitting by regularization. This paper addresses the above issues by 1) interpreting that the conventional training methods with regularization by noise injection optimize the lower bound of the true objective and 2) proposing a technique to achieve a tighter lower bound using multiple noise samples per training example in a stochastic gradient descent iteration. We demonstrate the effectiveness of our idea in several computer vision applications.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Feature-Flow Interpretation of Deep Convolutional Neural Networks
    Cui, Xinrui
    Wang, Dan
    Wang, Z. Jane
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) : 1847 - 1861
  • [42] Regularizing Deep Networks by Modeling and Predicting Label Structure
    Mostajabi, Mohammadreza
    Maire, Michael
    Shakhnarovich, Gregory
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 5629 - 5638
  • [43] JOINT OPTIMIZATION OF AUDIBLE NOISE SUPPRESSION AND DEEP NEURAL NETWORKS FOR SINGLE-CHANNEL SPEECH ENHANCEMENT
    Han, Wei
    Zhang, Xiongwei
    Min, Gang
    Sun, Meng
    Yang, Jibin
    2016 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO (ICME), 2016,
  • [44] Regularizing Neural Networks via Adversarial Model Perturbation
    Zheng, Yaowei
    Zhang, Richong
    Mao, Yongyi
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8152 - 8161
  • [45] Revisiting spatial dropout for regularizing convolutional neural networks
    Sanghun Lee
    Chulhee Lee
    Multimedia Tools and Applications, 2020, 79 : 34195 - 34207
  • [46] Revisiting spatial dropout for regularizing convolutional neural networks
    Lee, Sanghun
    Lee, Chulhee
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (45-46) : 34195 - 34207
  • [47] REGULARIZING NEURAL NETWORKS BY STOCHASTICALLY TRAINING LAYER ENSEMBLES
    Labach, Alex
    Valaee, Shahrokh
    PROCEEDINGS OF THE 2020 IEEE 30TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2020,
  • [48] Regularizing Neural Networks via Minimizing Hyperspherical Energy
    Lin, Rongmei
    Liu, Weiyang
    Liu, Zhen
    Feng, Chen
    Yu, Zhiding
    Rehg, James M.
    Xiong, Li
    Song, Le
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 6916 - 6925
  • [49] Regularizing Neural Networks via Stochastic Branch Layers
    Park, Wonpyo
    Seo, Paul Hongsuck
    Han, Bohyung
    Cho, Minsu
    ASIAN CONFERENCE ON MACHINE LEARNING, VOL 101, 2019, 101 : 678 - 693
  • [50] Guided Random Mask: Adaptively Regularizing Deep Neural Networks for Medical Image Analysis by Potential Lesions
    Yu, Xiaorui
    Wang, Shuqi
    Hu, Junjie
    APPLIED SCIENCES-BASEL, 2022, 12 (18):