Instance Enhancement Batch Normalization: An Adaptive Regulator of Batch Noise

被引:0
|
作者
Liang, Senwei [1 ]
Huang, Zhongzhan [2 ]
Liang, Mingfu [3 ]
Yang, Haizhao [1 ,4 ]
机构
[1] Purdue Univ, W Lafayette, IN 47907 USA
[2] Tsinghua Univ, Beijing, Peoples R China
[3] Northwestern Univ, Evanston, IL 60208 USA
[4] Natl Univ Singapore, Singapore, Singapore
来源
THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2020年 / 34卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Batch Normalization (BN) (Ioffe and Szegedy 2015) normalizes the features of an input image via statistics of a batch of images and hence BN will bring the noise to the gradient of training loss. Previous works indicate that the noise is important for the optimization and generalization of deep neural networks, but too much noise will harm the performance of networks. In our paper, we offer a new point of view that the self-attention mechanism can help to regulate the noise by enhancing instance-specific information to obtain a better regularization effect. Therefore, we propose an attention-based BN called Instance Enhancement Batch Normalization (IEBN) that recalibrates the information of each channel by a simple linear transformation. IEBN has a good capacity of regulating the batch noise and stabilizing network training to improve generalization even in the presence of two kinds of noise attacks during training. Finally, IEBN outperforms BN with only a light parameter increment in image classification tasks under different network structures and benchmark datasets.
引用
收藏
页码:4819 / 4827
页数:9
相关论文
共 50 条
  • [41] Accelerating Training of Batch Normalization: A Manifold Perspective
    Yi, Mingyang
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 180, 2022, 180 : 1128 - 1137
  • [42] Uncertainty Estimation via Stochastic Batch Normalization
    Atanov, Andrei
    Ashukha, Arsenii
    Molchanov, Dmitry
    Neklyudov, Kirill
    Vetrov, Dmitry
    ADVANCES IN NEURAL NETWORKS - ISNN 2019, PT I, 2019, 11554 : 261 - 269
  • [43] Removing Batch Normalization Boosts Adversarial Training
    Wang, Haotao
    Zhang, Aston
    Zheng, Shuai
    Shi, Xingjian
    Li, Mu
    Wang, Zhangyang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [44] The Unreasonable Effectiveness of the Final Batch Normalization Layer
    Kocaman, Veysel
    Shir, Ofer M.
    Back, Thomas
    ADVANCES IN VISUAL COMPUTING (ISVC 2021), PT II, 2021, 13018 : 81 - 93
  • [45] Impact of Batch Normalization on Convolutional Network Representations
    Potgieter, Hermanus L.
    Mouton, Coenraad
    Davel, Marelie H.
    ARTIFICIAL INTELLIGENCE RESEARCH, SACAIR 2024, 2025, 2326 : 235 - 252
  • [46] Double Forward Propagation for Memorized Batch Normalization
    Guo, Yong
    Wu, Qingyao
    Deng, Chaorui
    Chen, Jian
    Tan, Mingkui
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3134 - 3141
  • [47] BNET: Batch Normalization With Enhanced Linear Transformation
    Xu, Yuhui
    Xie, Lingxi
    Xie, Cihang
    Dai, Wenrui
    Mei, Jieru
    Qiao, Siyuan
    Shen, Wei
    Xiong, Hongkai
    Yuille, Alan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (07) : 9225 - 9232
  • [48] Batch Normalization Preconditioning for Neural Network Training
    Lange, Susanna
    Helfrich, Kyle
    Ye, Qiang
    Journal of Machine Learning Research, 2022, 23 : 1 - 41
  • [49] Representative Batch Normalization for Scene Text Recognition
    Sun, Yajie
    Cao, Xiaoling
    Sun, Yingying
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2022, 16 (07): : 2390 - 2406
  • [50] Revisiting Internal Covariate Shift for Batch Normalization
    Awais, Muhammad
    Bin Iqbal, Md Tauhid
    Bae, Sung-Ho
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (11) : 5082 - 5092