Wide Hidden Expansion Layer for Deep Convolutional Neural Networks

被引:0
|
作者
Wang, Min [1 ]
Liu, Baoyuan [2 ]
Foroosh, Hassan [1 ]
机构
[1] Univ Cent Florida, Orlando, FL 32816 USA
[2] Amazon, Seattle, WA USA
关键词
D O I
10.1109/wacv45572.2020.9093436
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Non-linearity is an essential factor contributing to the success of deep convolutional neural networks. Increasing the non-linearity in the network will enhance the network's learning capability, attributing to better performance. We present a novel Wide Hidden Expansion (WHE) layer that can significantly increase (by an order of magnitude) the number of activation functions in the network, with very little increase of computational complexity and memory consumption. It can be flexibly embedded with different network architectures to boost the performance of the original networks. The WHE layer is composed of a wide hidden layer, in which each channel only connects with two input channels and one output channel. Before connecting to the output channel, each intermediate channel in the WHE layer is followed by one activation function. In this manner, the number of activation functions can grow along with the number of channels in the hidden layer. We apply the WHE layer to ResNet, WideResNet, SENet, and MobileNet architectures and evaluate on ImageNet, CIFAR-100, and Tiny ImageNet dataset. On the ImageNet dataset, models with the WHE layer can achieve up to 2.01% higher Top-1 accuracy than baseline models, with less than 4% computation increase and less than 2% more parameters. On CIFAR-100 and Tiny ImageNet, when applying the WHE layer to ResNet models, it demonstrates consistent improvement in the accuracy of the networks. Applying the WHE layer to ResNet backbone of the CenterNet object detection model can also boost its performance on COCO and Pascal VOC datasets.
引用
收藏
页码:923 / 931
页数:9
相关论文
共 50 条
  • [1] Deep Convolutional Neural Networks with Layer-wise Context Expansion and Attention
    Yu, Dong
    Xiong, Wayne
    Droppo, Jasha
    Stolcke, Andreas
    Ye, Guoli
    Li, Jinyu
    Zweig, Geoffrey
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 17 - 21
  • [2] Hidden Layer Visualization for Convolutional Neural Networks: A Brief Review
    Rivera, Fabian
    Hurtado, Remigio
    PROCEEDINGS OF NINTH INTERNATIONAL CONGRESS ON INFORMATION AND COMMUNICATION TECHNOLOGY, ICICT 2024, VOL 3, 2024, 1013 : 471 - 482
  • [3] Modular Expansion of the Hidden Layer in Single Layer Feedforward Neural Networks
    Tissera, Migel D.
    McDonnell, Mark D.
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 2939 - 2945
  • [4] COMO: Efficient Deep Neural Networks Expansion With COnvolutional MaxOut
    Zhao, Baoxin
    Xiong, Haoyi
    Bian, Jiang
    Guo, Zhishan
    Xu, Cheng-Zhong
    Dou, Dejing
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 : 1722 - 1730
  • [5] Layer Removal for Transfer Learning with Deep Convolutional Neural Networks
    Zhi, Weiming
    Chen, Zhenghao
    Yueng, Henry Wing Fung
    Lu, Zhicheng
    Zandavi, Seid Miad
    Chung, Yuk Ying
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 460 - 469
  • [6] Predicting developed land expansion using deep convolutional neural networks
    Pourmohammadi, P.
    Adjeroh, D. A.
    Strager, M. P.
    Farid, Y. Z.
    ENVIRONMENTAL MODELLING & SOFTWARE, 2020, 134
  • [7] Inspecting Contraband Hidden in Milk Powder Using Deep Convolutional Neural Networks
    Sun, Xiaofei
    Pan, Wenwen
    Wang, Lei
    Wang, Xia
    Yang, Bin
    PROCEEDINGS OF 2017 IEEE 2ND INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC), 2017, : 1797 - 1801
  • [8] Deep Convolutional Neural Networks
    Gonzalez, Rafael C.
    IEEE SIGNAL PROCESSING MAGAZINE, 2018, 35 (06) : 79 - 87
  • [9] Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks
    Cao, Yuan
    Gu, Quanquan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [10] Deep Versus Wide Convolutional Neural Networks for Object Recognition on Neuromorphic System
    Alom, Md Zahangir
    Josue, Theodore
    Rahman, Md Nayim
    Mitchell, Will
    Yakopcic, Chris
    Taha, Tarek M.
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,