Analysis of non-linear activation functions for classification tasks using convolutional neural networks

被引:9
|
作者
Dureja A. [1 ]
Pahwa P. [2 ]
机构
[1] Computer Science & Engineering, USICT, GGSIPU, New Delhi
[2] Computer Science & Engineering, BPIT, Rohini, New Delhi
关键词
Activation function; CNN; Deep neural networks; Hidden layers; Machine learning; Non-linear problems;
D O I
10.2174/2213275911666181025143029
中图分类号
学科分类号
摘要
Background: In making the deep neural network, activation functions play an important role. But the choice of activation functions also affects the network in term of optimization and to retrieve the better results. Several activation functions have been introduced in machine learning for many practical applications. But which activation function should use at hidden layer of deep neural networks was not identified. Objective: The primary objective of this analysis was to describe which activation function must be used at hidden layers for deep neural networks to solve complex non-linear problems. Methods: The configuration for this comparative model was used by using the datasets of 2 classes (Cat/Dog). The number of Convolutional layer used in this network was 3 and the pooling layer was also introduced after each layer of CNN layer. The total of the dataset was divided into the two parts. The first 8000 images were mainly used for training the network and the next 2000 images were used for testing the network. Results: The experimental comparison was done by analyzing the network by taking different activation functions on each layer of CNN network. The validation error and accuracy on Cat/Dog dataset were analyzed using activation functions (ReLU, Tanh, Selu, PRelu, Elu) at number of hidden layers. Overall the Relu gave best performance with the validation loss at 25th Epoch 0.3912 and validation accuracy at 25th Epoch 0.8320. Conclusion: It is found that a CNN model with ReLU hidden layers (3 hidden layers here) gives best results and improve overall performance better in term of accuracy and speed. These advantages of ReLU in CNN at number of hidden layers are helpful to effectively and fast retrieval of images from the databases. © 2019 Bentham Science Publishers.
引用
收藏
页码:156 / 161
页数:5
相关论文
共 50 条
  • [1] ECG Arrhythmia Classification Using Non-Linear Features and Convolutional Neural Networks
    Cajas, Sebastian
    Astaiza, Pedro
    Garcia-Chicangana, David Santiago
    Segura, Camilo
    Lopez, Diego M.
    2020 COMPUTING IN CARDIOLOGY, 2020,
  • [2] Verification of LSTM Neural Networks with Non-linear Activation Functions
    Moradkhani, Farzaneh
    Fibich, Connor
    Franzle, Martin
    NASA FORMAL METHODS, NFM 2023, 2023, 13903 : 1 - 15
  • [3] Non-linear survival analysis using neural networks
    Ripley, RM
    Harris, AL
    Tarassenko, L
    STATISTICS IN MEDICINE, 2004, 23 (05) : 825 - 842
  • [4] NISQ-Friendly Non-Linear Activation Functions for Quantum Neural Networks
    Sajadimanesh, Sohrab
    Faye, Jean Paul Latyr
    Atoofian, Ehsan
    2022 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, ARCHITECTURE AND STORAGE (NAS), 2022, : 121 - 128
  • [5] Neural networks with orthonormal activation functions for non-linear dynamic system identification
    Shukla, Deepak
    Paul, Frank W.
    International Journal of Smart Engineering System Design, 1999, 2 (02): : 79 - 88
  • [6] Learning continuous piecewise non-linear activation functions for deep neural networks
    Gao, Xinchen
    Li, Yawei
    Li, Wen
    Duan, Lixin
    Van Gool, Luc
    Benini, Luca
    Magno, Michele
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1835 - 1840
  • [7] On the application of neural networks to non-linear image processing tasks
    de Ridder, D
    Duin, RPW
    Verbeek, PW
    van Vliet, LJ
    ICONIP'98: THE FIFTH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING JOINTLY WITH JNNS'98: THE 1998 ANNUAL CONFERENCE OF THE JAPANESE NEURAL NETWORK SOCIETY - PROCEEDINGS, VOLS 1-3, 1998, : 161 - 165
  • [8] Classification of Linear Processes Type Using Convolutional Neural Networks
    Kalantari, S. H.
    Kalhor, A.
    Araabi, B. N.
    2022 8TH INTERNATIONAL CONFERENCE ON CONTROL, DECISION AND INFORMATION TECHNOLOGIES (CODIT'22), 2022, : 914 - 919
  • [9] A new learning algorithm for neural networks with integer weights and quantized non-linear activation functions
    Yan Yi
    Zhang Hangping
    Zhou Bin
    ARTIFICIAL INTELLIGENCE IN THEORY AND PRACTICE II, 2008, 276 : 427 - 431
  • [10] A Monte Carlo Simulation Approach in Non-linear Structural Dynamics Using Convolutional Neural Networks
    Bamer, Franz
    Thaler, Denny
    Stoffel, Marcus
    Markert, Bernd
    FRONTIERS IN BUILT ENVIRONMENT, 2021, 7