Hyperbolic Linear Units for Deep Convolutional Neural Networks

被引:0
|
作者
Li, Jia [1 ]
Xu, Hua [1 ]
Deng, Junhui [1 ]
Sun, Xiaomin [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Tsinghua Natl Lab Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, rectified linear units (ReLUs) have been used to solve the vanishing gradient problem. Their use has led to state-of-the-art results in various problems such as image classification. In this paper, we propose the hyperbolic linear units (HLUs) which not only speed up learning process in deep convolutional neural networks but also obtain better performance in image classification tasks. Unlike ReLUs, HLUs have inheriently negative values which could make mean unit outputs closer to zero. Mean unit outputs close to zero means we can speed up the learning process because they bring the normal gradient close to the natural gradient. Indeed, the difference called bias shift between natural gradient and the normal gradient is related to the mean activation of input units. Experiments with three popular CNN architectures, LeNet, Inception network and ResNet on various benchmarks including MNIST, CIFAR-10 and CIFAR-100 demonstrate that our proposed HLUs achieve significant improvement compared to other commonly used activation functions.
引用
收藏
页码:353 / 359
页数:7
相关论文
共 50 条
  • [1] Elastic exponential linear units for convolutional neural networks
    Kim, Daeho
    Kim, Jinah
    Kim, Jaeil
    NEUROCOMPUTING, 2020, 406 : 253 - 266
  • [2] Improving deep convolutional neural networks with mixed maxout units
    Zhao, Hui-zhen
    Liu, Fu-xian
    Li, Long-yue
    PLOS ONE, 2017, 12 (07):
  • [3] Linear Regularized Compression of Deep Convolutional Neural Networks
    Ceruti, Claudio
    Campadelli, Paola
    Casiraghi, Elena
    IMAGE ANALYSIS AND PROCESSING,(ICIAP 2017), PT I, 2017, 10484 : 244 - 253
  • [4] Fully hyperbolic convolutional neural networks
    Keegan Lensink
    Bas Peters
    Eldad Haber
    Research in the Mathematical Sciences, 2022, 9
  • [5] Fully hyperbolic convolutional neural networks
    Lensink, Keegan
    Peters, Bas
    Haber, Eldad
    RESEARCH IN THE MATHEMATICAL SCIENCES, 2022, 9 (04)
  • [6] Hyperbolic Graph Convolutional Neural Networks
    Chami, Ines
    Ying, Rex
    Re, Christopher
    Leskovec, Jure
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [7] FReLU: Flexible Rectified Linear Units for Improving Convolutional Neural Networks
    Qiu, Suo
    Xu, Xiangmin
    Cai, Bolun
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 1223 - 1228
  • [8] Parametric Exponential Linear Unit for Deep Convolutional Neural Networks
    Trottier, Ludovic
    Giguere, Philippe
    Chaib-draa, Brahim
    2017 16TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2017, : 207 - 214
  • [9] A novel softplus linear unit for deep convolutional neural networks
    Huizhen Zhao
    Fuxian Liu
    Longyue Li
    Chang Luo
    Applied Intelligence, 2018, 48 : 1707 - 1720
  • [10] A novel softplus linear unit for deep convolutional neural networks
    Zhao, Huizhen
    Liu, Fuxian
    Li, Longyue
    Luo, Chang
    APPLIED INTELLIGENCE, 2018, 48 (07) : 1707 - 1720