The Weights Reset Technique for Deep Neural Networks Implicit Regularization

被引:0
|
作者
Plusch, Grigoriy [1 ]
Arsenyev-Obraztsov, Sergey [1 ]
Kochueva, Olga [1 ]
机构
[1] Natl Univ Oil & Gas Gubkin Univ, Dept Appl Math & Comp Modeling, 65 Leninsky Prospekt, Moscow 119991, Russia
关键词
machine learning; deep learning; implicit regularization; computer vision; REPRESENTATIONS;
D O I
10.3390/computation11080148
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
We present a new regularization method called Weights Reset, which includes periodically resetting a random portion of layer weights during the training process using predefined probability distributions. This technique was applied and tested on several popular classification datasets, Caltech-101, CIFAR-100 and Imagenette. We compare these results with other traditional regularization methods. The subsequent test results demonstrate that the Weights Reset method is competitive, achieving the best performance on Imagenette dataset and the challenging and unbalanced Caltech-101 dataset. This method also has sufficient potential to prevent vanishing and exploding gradients. However, this analysis is of a brief nature. Further comprehensive studies are needed in order to gain a deep understanding of the computing potential and limitations of the Weights Reset method. The observed results show that the Weights Reset method can be estimated as an effective extension of the traditional regularization methods and can help to improve model performance and generalization.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Implicit regularization of a deep augmented neural network model for human motion prediction
    Yadav, Gaurav Kumar
    Abdel-Nasser, Mohamed
    Rashwan, Hatem A.
    Puig, Domenec
    Nandi, G. C.
    APPLIED INTELLIGENCE, 2023, 53 (14) : 18027 - 18040
  • [32] Implicit Regularization in Deep Matrix Factorization
    Arora, Sanjeev
    Cohen, Nadav
    Hu, Wei
    Luo, Yuping
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [33] Implicit Regularization in Deep Tensor Factorization
    Milanesi, Paolo
    Kadri, Hachem
    Ayache, Stephan
    Artieres, Thierry
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [34] ABOUT AN ALGORITHM FOR CONSISTENT WEIGHTS INITIALIZATION OF DEEP NEURAL NETWORKS AND NEURAL NETWORKS ENSEMBLE LEARNING
    Drokin, I. S.
    VESTNIK SANKT-PETERBURGSKOGO UNIVERSITETA SERIYA 10 PRIKLADNAYA MATEMATIKA INFORMATIKA PROTSESSY UPRAVLENIYA, 2016, 12 (04): : 66 - 74
  • [35] Deep Neural Networks Pruning via the Structured Perspective Regularization
    Cacciola, Matteo
    Frangioni, Antonio
    Li, Xinlin
    Lodi, Andrea
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2023, 5 (04): : 1051 - 1077
  • [36] Deep Spiking Neural Networks With Binary Weights for Object Recognition
    Wang, Yixuan
    Xu, Yang
    Yan, Rui
    Tang, Huajin
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2021, 13 (03) : 514 - 523
  • [37] Hierarchical Group Sparse Regularization for Deep Convolutional Neural Networks
    Mitsuno, Kakeru
    Miyao, Junichi
    Kurita, Takio
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [38] The Analysis of Regularization in Deep Neural Networks Using Metagraph Approach
    Fedorenko, Yuriy S.
    Gapanyuk, Yuriy E.
    Minakova, Svetlana V.
    ADVANCES IN NEURAL COMPUTATION, MACHINE LEARNING, AND COGNITIVE RESEARCH, 2018, 736 : 3 - 8
  • [39] Pipelined Training with Stale Weights in Deep Convolutional Neural Networks
    Zhang, Lifu
    Abdelrahman, Tarek S.
    APPLIED COMPUTATIONAL INTELLIGENCE AND SOFT COMPUTING, 2021, 2021
  • [40] A Hardware Accelerator Based on Quantized Weights for Deep Neural Networks
    Sreehari, R.
    Deepu, Vijayasenan
    Arulalan, M. R.
    EMERGING RESEARCH IN ELECTRONICS, COMPUTER SCIENCE AND TECHNOLOGY, ICERECT 2018, 2019, 545 : 1079 - 1091