Instance Enhancement Batch Normalization: An Adaptive Regulator of Batch Noise

被引:0
|
作者
Liang, Senwei [1 ]
Huang, Zhongzhan [2 ]
Liang, Mingfu [3 ]
Yang, Haizhao [1 ,4 ]
机构
[1] Purdue Univ, W Lafayette, IN 47907 USA
[2] Tsinghua Univ, Beijing, Peoples R China
[3] Northwestern Univ, Evanston, IL 60208 USA
[4] Natl Univ Singapore, Singapore, Singapore
来源
THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2020年 / 34卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Batch Normalization (BN) (Ioffe and Szegedy 2015) normalizes the features of an input image via statistics of a batch of images and hence BN will bring the noise to the gradient of training loss. Previous works indicate that the noise is important for the optimization and generalization of deep neural networks, but too much noise will harm the performance of networks. In our paper, we offer a new point of view that the self-attention mechanism can help to regulate the noise by enhancing instance-specific information to obtain a better regularization effect. Therefore, we propose an attention-based BN called Instance Enhancement Batch Normalization (IEBN) that recalibrates the information of each channel by a simple linear transformation. IEBN has a good capacity of regulating the batch noise and stabilizing network training to improve generalization even in the presence of two kinds of noise attacks during training. Finally, IEBN outperforms BN with only a light parameter increment in image classification tasks under different network structures and benchmark datasets.
引用
收藏
页码:4819 / 4827
页数:9
相关论文
共 50 条
  • [21] Enhanced LSTM with Batch Normalization
    Wang, Li-Na
    Zhong, Guoqiang
    Yan, Shoujun
    Dong, Junyu
    Huang, Kaizhu
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 746 - 755
  • [22] Riemannian approach to batch normalization
    Cho, Minhyung
    Lee, Jaehyung
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [23] Batch Normalization and Dropout Regularization in Training Deep Neural Networks with Label Noise
    Rusiecki, Andrzej
    INTELLIGENT SYSTEMS DESIGN AND APPLICATIONS, ISDA 2021, 2022, 418 : 57 - 66
  • [24] Cross-Iteration Batch Normalization
    Yao, Zhuliang
    Cao, Yue
    Zheng, Shuxin
    Huang, Gao
    Lin, Stephen
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 12326 - 12335
  • [25] Data-Free Quantization with Accurate Activation Clipping and Adaptive Batch Normalization
    Yefei He
    Luoming Zhang
    Weijia Wu
    Hong Zhou
    Neural Processing Letters, 2023, 55 : 10555 - 10568
  • [26] Data-Free Quantization with Accurate Activation Clipping and Adaptive Batch Normalization
    He, Yefei
    Zhang, Luoming
    Wu, Weijia
    Zhou, Hong
    NEURAL PROCESSING LETTERS, 2023, 55 (08) : 10555 - 10568
  • [27] Leveraging Batch Normalization for Vision Transformers
    Yao, Zhuliang
    Cao, Yue
    Lin, Yutong
    Liu, Ze
    Zhang, Zheng
    Hu, Han
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 413 - 422
  • [28] Representative Batch Normalization with Feature Calibration
    Gao, Shang-Hua
    Han, Qi
    Li, Duo
    Cheng, Ming-Ming
    Peng, Pai
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8665 - 8675
  • [29] The Effect of Batch Normalization in the Symmetric Phase
    Takagi, Shiro
    Yoshida, Yuki
    Okada, Masato
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 229 - 240
  • [30] Proxy-Normalizing Activations to Match Batch Normalization while Removing Batch Dependence
    Labatie, Antoine
    Masters, Dominic
    Eaton-Rosen, Zach
    Luschi, Carlo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34