αSechSig and αTanhSig: two novel non-monotonic activation functions

被引:4
|
作者
Kozkurt, Cemil [1 ]
Kilicarslan, Serhat [2 ]
Bas, Selcuk [3 ]
Elen, Abdullah [2 ]
机构
[1] Bandirma Onyedi Eylul Univ, Fac Engn & Nat Sci, Dept Comp Engn, Dept Transportat Engn, Bandirmabalikesir, Turkiye
[2] Bandirma Onyedi Eylul Univ, Fac Engn & Nat Sci, Dept Software Engn, TR-10200 Bandirmabalikesir, Turkiye
[3] Bandirma Onyedi Eylul Univ, Dept Accounting & Tax, Bandirma Vocat Sch, Bandirmabalikesir, Turkiye
关键词
Activation function; Deep learning; alpha SechSig and alpha TanhSig activation functions;
D O I
10.1007/s00500-023-09279-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The deep learning architectures' activation functions play a significant role in processing the data entering the network to provide the most appropriate output. Activation functions (AF) are created by taking into consideration aspects like avoiding model local minima and improving training efficiency. Negative weights and vanishing gradients are frequently taken into account by the AF suggested in the literature. Recently, a number of non-monotonic AF have increasingly replaced previous methods for improving convolutional neural network (CNN) performance. In this study, two novel non-linear non-monotonic activation functions, alpha-SechSig and alpha-TanhSig are proposed that can overcome the existing problems. The negative part of alpha-SechSig and alpha-TanhSig is non-monotonic and approaches zero as the negative input decreases, allowing the negative part to retain its sparsity while introducing negative activation values and non-zero derivative values. In experimental evaluations, alpha-SechSig and alpha-TanhSig activation functions were tested on MNIST, KMNIST, Svhn_Cropped, STL-10, and CIFAR-10 datasets. In addition, better results were obtained than the non-monotonic Swish, Logish, Mish, Smish, and monotonic ReLU, SinLU, and LReLU AF known in the literature. Moreover, the best accuracy score for the alpha SechSig and alpha TanhSig activation functions was obtained with MNIST at 0.9959 and 0.9956, respectively.
引用
收藏
页码:18451 / 18467
页数:17
相关论文
共 50 条
  • [41] COMPLEXITY OF NON-MONOTONIC LOGICS
    Toran, Jacob
    Thomas, Michael
    Vollmer, Heribert
    BULLETIN OF THE EUROPEAN ASSOCIATION FOR THEORETICAL COMPUTER SCIENCE, 2010, (102): : 53 - 82
  • [42] The Non-Monotonic NonFregean Logic
    Wojtowicz, Anna
    FILOZOFIA NAUKI, 2011, 19 (02): : 105 - +
  • [43] Non-monotonic Reasoning on the Web
    Cristani, Matteo
    WEB ENGINEERING (ICWE 2019), 2019, 11496 : 586 - 589
  • [44] Non-monotonic Temporal Goals
    Goncalves, Ricardo
    Knorr, Matthias
    Leite, Joao
    Slota, Martin
    LOGIC PROGRAMMING AND NONMONOTONIC REASONING (LPNMR 2013), 2013, 8148 : 374 - 386
  • [45] Tabling for non-monotonic programming
    Swift, T
    ANNALS OF MATHEMATICS AND ARTIFICIAL INTELLIGENCE, 1999, 25 (3-4) : 201 - 240
  • [46] On non-monotonic strategic reasoning
    Catonini, Emiliano
    GAMES AND ECONOMIC BEHAVIOR, 2020, 120 : 209 - 224
  • [47] Stability of uncertain systems using Lyapunov functions with non-monotonic terms
    Lacerda, Marcio J.
    Seiler, Peter
    AUTOMATICA, 2017, 82 : 187 - 193
  • [48] Computation of non-monotonic Lyapunov functions for continuous-time systems
    Li, Huijuan
    Liu, AnPing
    COMMUNICATIONS IN NONLINEAR SCIENCE AND NUMERICAL SIMULATION, 2017, 50 : 35 - 50
  • [49] Non-monotonic magnetoresistivity in two-dimensional electron systems
    Wang, Yi-Ting
    Woo, Tak-Pong
    Lo, Shun-Tsung
    Kim, Gil-Ho
    Liang, Chi-Te
    JOURNAL OF THE KOREAN PHYSICAL SOCIETY, 2014, 65 (10) : 1503 - 1507
  • [50] Using gradual numbers to analyze non-monotonic functions of fuzzy intervals
    Untiedt, Elizabeth
    Lodwick, Weldon
    2008 ANNUAL MEETING OF THE NORTH AMERICAN FUZZY INFORMATION PROCESSING SOCIETY, VOLS 1 AND 2, 2008, : 590 - 595