The Weights Reset Technique for Deep Neural Networks Implicit Regularization

被引:0
|
作者
Plusch, Grigoriy [1 ]
Arsenyev-Obraztsov, Sergey [1 ]
Kochueva, Olga [1 ]
机构
[1] Natl Univ Oil & Gas Gubkin Univ, Dept Appl Math & Comp Modeling, 65 Leninsky Prospekt, Moscow 119991, Russia
关键词
machine learning; deep learning; implicit regularization; computer vision; REPRESENTATIONS;
D O I
10.3390/computation11080148
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
We present a new regularization method called Weights Reset, which includes periodically resetting a random portion of layer weights during the training process using predefined probability distributions. This technique was applied and tested on several popular classification datasets, Caltech-101, CIFAR-100 and Imagenette. We compare these results with other traditional regularization methods. The subsequent test results demonstrate that the Weights Reset method is competitive, achieving the best performance on Imagenette dataset and the challenging and unbalanced Caltech-101 dataset. This method also has sufficient potential to prevent vanishing and exploding gradients. However, this analysis is of a brief nature. Further comprehensive studies are needed in order to gain a deep understanding of the computing potential and limitations of the Weights Reset method. The observed results show that the Weights Reset method can be estimated as an effective extension of the traditional regularization methods and can help to improve model performance and generalization.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] GradAug: A New Regularization Method for Deep Neural Networks
    Yang, Taojiannan
    Zhu, Sijie
    Chen, Chen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [22] Towards Robustness of Deep Neural Networks via Regularization
    Li, Yao
    Min, Martin Renqiang
    Lee, Thomas
    Yu, Wenchao
    Kruus, Erik
    Wang, Wei
    Hsieh, Cho-Jui
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7476 - 7485
  • [23] Bridgeout: Stochastic Bridge Regularization for Deep Neural Networks
    Khan, Najeeb
    Shah, Jawad
    Stavness, Ian
    IEEE ACCESS, 2018, 6 : 42961 - 42970
  • [24] Adaptive Knowledge Driven Regularization for Deep Neural Networks
    Luo, Zhaojing
    Cai, Shaofeng
    Cui, Can
    Ooi, Beng Chin
    Yang, Yang
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8810 - 8818
  • [25] Regional Tree Regularization for Interpretability in Deep Neural Networks
    Wu, Mike
    Parbhoo, Sonali
    Hughes, Michael C.
    Kindle, Ryan
    Celi, Leo
    Zazzi, Maurizio
    Roth, Volker
    Doshi-Velez, Finale
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 6413 - 6421
  • [26] On the Security Relevance of Initial Weights in Deep Neural Networks
    Grosse, Kathrin
    Trost, Thomas A.
    Mosbach, Marius
    Backes, Michael
    Klakow, Dietrich
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT I, 2020, 12396 : 3 - 14
  • [27] Optimizing for interpretability in deep neural networks with tree regularization
    Wu M.
    Parbhoo S.
    Hughes M.C.
    Roth V.
    Doshi-Velez F.
    Journal of Artificial Intelligence Research, 2021, 72
  • [28] Theory of adaptive SVD regularization for deep neural networks
    Bejani, Mohammad Mahdi
    Ghatee, Mehdi
    NEURAL NETWORKS, 2020, 128 : 33 - 46
  • [29] Optimizing for Interpretability in Deep Neural Networks with Tree Regularization
    Wu, Mike
    Parbhoo, Sonali
    Hughes, Michael C.
    Roth, Volker
    Doshi-Velez, Finale
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2021, 72 : 1 - 37
  • [30] Generalize Deep Neural Networks With Adaptive Regularization for Classifying
    Guo, Kehua
    Tao, Ze
    Zhang, Lingyan
    Hu, Bin
    Kui, Xiaoyan
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (01) : 1216 - 1229