αSechSig and αTanhSig: two novel non-monotonic activation functions

被引:4
|
作者
Kozkurt, Cemil [1 ]
Kilicarslan, Serhat [2 ]
Bas, Selcuk [3 ]
Elen, Abdullah [2 ]
机构
[1] Bandirma Onyedi Eylul Univ, Fac Engn & Nat Sci, Dept Comp Engn, Dept Transportat Engn, Bandirmabalikesir, Turkiye
[2] Bandirma Onyedi Eylul Univ, Fac Engn & Nat Sci, Dept Software Engn, TR-10200 Bandirmabalikesir, Turkiye
[3] Bandirma Onyedi Eylul Univ, Dept Accounting & Tax, Bandirma Vocat Sch, Bandirmabalikesir, Turkiye
关键词
Activation function; Deep learning; alpha SechSig and alpha TanhSig activation functions;
D O I
10.1007/s00500-023-09279-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The deep learning architectures' activation functions play a significant role in processing the data entering the network to provide the most appropriate output. Activation functions (AF) are created by taking into consideration aspects like avoiding model local minima and improving training efficiency. Negative weights and vanishing gradients are frequently taken into account by the AF suggested in the literature. Recently, a number of non-monotonic AF have increasingly replaced previous methods for improving convolutional neural network (CNN) performance. In this study, two novel non-linear non-monotonic activation functions, alpha-SechSig and alpha-TanhSig are proposed that can overcome the existing problems. The negative part of alpha-SechSig and alpha-TanhSig is non-monotonic and approaches zero as the negative input decreases, allowing the negative part to retain its sparsity while introducing negative activation values and non-zero derivative values. In experimental evaluations, alpha-SechSig and alpha-TanhSig activation functions were tested on MNIST, KMNIST, Svhn_Cropped, STL-10, and CIFAR-10 datasets. In addition, better results were obtained than the non-monotonic Swish, Logish, Mish, Smish, and monotonic ReLU, SinLU, and LReLU AF known in the literature. Moreover, the best accuracy score for the alpha SechSig and alpha TanhSig activation functions was obtained with MNIST at 0.9959 and 0.9956, respectively.
引用
收藏
页码:18451 / 18467
页数:17
相关论文
共 50 条
  • [1] α­SechSig and α­TanhSig: two novel non-monotonic activation functions
    Cemil Közkurt
    Serhat Kiliçarslan
    Selçuk Baş
    Abdullah Elen
    Soft Computing, 2023, 27 : 18451 - 18467
  • [2] PFLU and FPFLU: Two novel non-monotonic activation functions in convolutional neural networks
    Zhu, Meng
    Min, Weidong
    Wang, Qi
    Zou, Song
    Chen, Xinhao
    NEUROCOMPUTING, 2021, 429 : 110 - 117
  • [3] Learning a Single Neuron for Non-monotonic Activation Functions
    Wu, Lei
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [4] Non-monotonic Explanation Functions
    Amgoud, Leila
    SYMBOLIC AND QUANTITATIVE APPROACHES TO REASONING WITH UNCERTAINTY, ECSQARU 2021, 2021, 12897 : 19 - 31
  • [5] ErfAct and Pserf: Non-monotonic Smooth Trainable Activation Functions
    Biswas, Koushik
    Kumar, Sandeep
    Banerjee, Shilpak
    Pandey, Ashish Kumar
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 6097 - 6105
  • [6] Supervised learning for multilayered neural network with non-monotonic activation functions
    Kotani, M
    Akazawa, K
    NEURAL NETWORKS FOR SIGNAL PROCESSING VI, 1996, : 13 - 22
  • [7] Multistability of Memristive Neural Networks with Non-monotonic Piecewise Linear Activation Functions
    Nie, Xiaobing
    Cao, Jinde
    ADVANCES IN NEURAL NETWORKS - ISNN 2015, 2015, 9377 : 182 - 191
  • [8] Global dynamics of discrete neural networks allowing non-monotonic activation functions
    Liz, Eduardo
    Ruiz-Herrera, Alfonso
    NONLINEARITY, 2014, 27 (02) : 289 - 304
  • [9] Multistability and Instability of Neural Networks with Non-monotonic Piecewise Linear Activation Functions
    Nie, Xiaobing
    Cao, Linde
    2015 SEVENTH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2015, : 152 - 157
  • [10] A fixed point theorem for non-monotonic functions
    Esik, Zoltan
    Rondogiannis, Panos
    THEORETICAL COMPUTER SCIENCE, 2015, 574 : 18 - 38