Wide Hidden Expansion Layer for Deep Convolutional Neural Networks

被引:0
|
作者
Wang, Min [1 ]
Liu, Baoyuan [2 ]
Foroosh, Hassan [1 ]
机构
[1] Univ Cent Florida, Orlando, FL 32816 USA
[2] Amazon, Seattle, WA USA
关键词
D O I
10.1109/wacv45572.2020.9093436
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Non-linearity is an essential factor contributing to the success of deep convolutional neural networks. Increasing the non-linearity in the network will enhance the network's learning capability, attributing to better performance. We present a novel Wide Hidden Expansion (WHE) layer that can significantly increase (by an order of magnitude) the number of activation functions in the network, with very little increase of computational complexity and memory consumption. It can be flexibly embedded with different network architectures to boost the performance of the original networks. The WHE layer is composed of a wide hidden layer, in which each channel only connects with two input channels and one output channel. Before connecting to the output channel, each intermediate channel in the WHE layer is followed by one activation function. In this manner, the number of activation functions can grow along with the number of channels in the hidden layer. We apply the WHE layer to ResNet, WideResNet, SENet, and MobileNet architectures and evaluate on ImageNet, CIFAR-100, and Tiny ImageNet dataset. On the ImageNet dataset, models with the WHE layer can achieve up to 2.01% higher Top-1 accuracy than baseline models, with less than 4% computation increase and less than 2% more parameters. On CIFAR-100 and Tiny ImageNet, when applying the WHE layer to ResNet models, it demonstrates consistent improvement in the accuracy of the networks. Applying the WHE layer to ResNet backbone of the CenterNet object detection model can also boost its performance on COCO and Pascal VOC datasets.
引用
收藏
页码:923 / 931
页数:9
相关论文
共 50 条
  • [31] Flattening Layer Pruning in Convolutional Neural Networks
    Jeczmionek, Ernest
    Kowalski, Piotr A.
    SYMMETRY-BASEL, 2021, 13 (07):
  • [32] Learning hidden chemistry with deep neural networks
    Nguyen, Tien-Cuong
    Nguyen, Van-Quyen
    Ngo, Van-Linh
    Than, Quang-Khoat
    Pham, Tien-Lam
    COMPUTATIONAL MATERIALS SCIENCE, 2021, 200
  • [33] Learning hidden elasticity with deep neural networks
    Chen, Chun-Teh
    Gu, Grace X.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2021, 118 (31)
  • [34] Discriminative Layer Pruning for Convolutional Neural Networks
    Jordao, Artur
    Lie, Maiko
    Schwartz, William Robson
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 828 - 837
  • [35] Elastography mapped by deep convolutional neural networks
    LIU DongXu
    KRUGGEL Frithjof
    SUN LiZhi
    Science China(Technological Sciences), 2021, (07) : 1567 - 1574
  • [36] Plug and Play Deep Convolutional Neural Networks
    Neary, Patrick
    Allan, Vicki
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 388 - 395
  • [37] An Efficient Accelerator for Deep Convolutional Neural Networks
    Kuo, Yi-Xian
    Lai, Yeong-Kang
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [38] Metaphase finding with deep convolutional neural networks
    Moazzen, Yaser
    Capar, Abdulkerim
    Albayrak, Abdulkadir
    Calik, Nurullah
    Toreyin, Behcet Ugur
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2019, 52 : 353 - 361
  • [39] Deep distributed convolutional neural networks: Universality
    Zhou, Ding-Xuan
    ANALYSIS AND APPLICATIONS, 2018, 16 (06) : 895 - 919
  • [40] Predicting enhancers with deep convolutional neural networks
    Min, Xu
    Zeng, Wanwen
    Chen, Shengquan
    Chen, Ning
    Chen, Ting
    Jiang, Rui
    BMC BIOINFORMATICS, 2017, 18