αSechSig and αTanhSig: two novel non-monotonic activation functions

被引:4
|
作者
Kozkurt, Cemil [1 ]
Kilicarslan, Serhat [2 ]
Bas, Selcuk [3 ]
Elen, Abdullah [2 ]
机构
[1] Bandirma Onyedi Eylul Univ, Fac Engn & Nat Sci, Dept Comp Engn, Dept Transportat Engn, Bandirmabalikesir, Turkiye
[2] Bandirma Onyedi Eylul Univ, Fac Engn & Nat Sci, Dept Software Engn, TR-10200 Bandirmabalikesir, Turkiye
[3] Bandirma Onyedi Eylul Univ, Dept Accounting & Tax, Bandirma Vocat Sch, Bandirmabalikesir, Turkiye
关键词
Activation function; Deep learning; alpha SechSig and alpha TanhSig activation functions;
D O I
10.1007/s00500-023-09279-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The deep learning architectures' activation functions play a significant role in processing the data entering the network to provide the most appropriate output. Activation functions (AF) are created by taking into consideration aspects like avoiding model local minima and improving training efficiency. Negative weights and vanishing gradients are frequently taken into account by the AF suggested in the literature. Recently, a number of non-monotonic AF have increasingly replaced previous methods for improving convolutional neural network (CNN) performance. In this study, two novel non-linear non-monotonic activation functions, alpha-SechSig and alpha-TanhSig are proposed that can overcome the existing problems. The negative part of alpha-SechSig and alpha-TanhSig is non-monotonic and approaches zero as the negative input decreases, allowing the negative part to retain its sparsity while introducing negative activation values and non-zero derivative values. In experimental evaluations, alpha-SechSig and alpha-TanhSig activation functions were tested on MNIST, KMNIST, Svhn_Cropped, STL-10, and CIFAR-10 datasets. In addition, better results were obtained than the non-monotonic Swish, Logish, Mish, Smish, and monotonic ReLU, SinLU, and LReLU AF known in the literature. Moreover, the best accuracy score for the alpha SechSig and alpha TanhSig activation functions was obtained with MNIST at 0.9959 and 0.9956, respectively.
引用
收藏
页码:18451 / 18467
页数:17
相关论文
共 50 条
  • [21] Identifying monotonic and non-monotonic relationships
    Yitzhaki, Shlomo
    Schechtman, Edna
    ECONOMICS LETTERS, 2012, 116 (01) : 23 - 25
  • [22] Monotonic and non-monotonic logics of knowledge
    Parikh, Rohit
    Fundamenta Mathematicae, 1991, 15 (3-4) : 255 - 274
  • [23] Monotonic and Non-monotonic Context Delegation
    AL-Wahah, Mouiad
    Farkas, Csilla
    PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON INFORMATION SYSTEMS SECURITY AND PRIVACY (ICISSP), 2019, : 449 - 460
  • [24] Non-monotonic negativity
    Nishiguchi, S
    PACLIC 17: LANGUAGE, INFORMATION AND COMPUTATION, PROCEEDINGS, 2003, : 204 - 215
  • [25] Two representation theorems for non-monotonic inference relations
    Zhu, Zhaohui
    Xiao, Wenjie
    JOURNAL OF LOGIC AND COMPUTATION, 2007, 17 (04) : 727 - 747
  • [26] A method for the reconstruction of unknown non-monotonic growth functions in the chemostat
    Jan Sieber
    Alain Rapaport
    Serafim Rodrigues
    Mathieu Desroches
    Bioprocess and Biosystems Engineering, 2013, 36 : 1497 - 1507
  • [27] A method for the reconstruction of unknown non-monotonic growth functions in the chemostat
    Sieber, Jan
    Rapaport, Alain
    Rodrigues, Serafim
    Desroches, Mathieu
    BIOPROCESS AND BIOSYSTEMS ENGINEERING, 2013, 36 (10) : 1497 - 1507
  • [28] Image Reduction Operators Based on Non-Monotonic Averaging Functions
    Wilkin, Tim
    2013 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ - IEEE 2013), 2013,
  • [29] Control with General Dissipativity Constraint and Non-Monotonic Lyapunov Functions
    Tri Tran
    Maciejowski, Jan
    2016 INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES (ICCAIS), 2016, : 177 - 182
  • [30] Multi-stack hybrid CNN with non-monotonic activation functions for hyperspectral satellite image classification
    Mainak Bandyopadhyay
    Neural Computing and Applications, 2021, 33 : 14809 - 14822