Training Strategies for Convolutional Neural Networks with Transformed Input

被引:0
|
作者
Khandani, Masoumeh Kalantari [1 ]
Mikhael, Wasfy B. [1 ]
机构
[1] Univ Cent Florida, Dept Elect & Comp Engn, Orlando, FL 32816 USA
关键词
image classification; Convolutional neural networks; DCT; DWT; domain transforms;
D O I
10.1109/MWSCAS47672.2021.9531913
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Network (CNN) are now considered the main tool for image classification. However, most networks studied for classification are large and have extensive computing and storage requirements. Their training time is also usually very long. Such costly computational and storage requirements cannot be met in many applications with simple devices such as small processors or Internet of things (IoTs) devices. Therefore, reducing the size of networks and input sizes become necessary. However, such reductions are not easy and may reduce the classification performance. We examine how domain transforms under different training strategies can be used for efficient size reduction and improvement of classification accuracy. In this paper, we consider networks with under 220K learnable parameters, as opposed to millions in deeper networks. We show that by representing the input to a CNN using appropriately selected domain transforms, such as discrete wavelet transforms (DWT) or discrete cosine transform (DCT), it is possible to efficiently improve the performance of size-reduced networks. For example, DWT proves to be very effective when significant size reduction is needed (improving the result by up to 9%). It is also shown that by tuning training strategies such as the number of epochs and mini batch size, the performance can be further improved by up to 4% under fixed training time.
引用
收藏
页码:1058 / 1061
页数:4
相关论文
共 50 条
  • [21] Using different training strategies for urban land-use classification based on convolutional neural networks
    Qiu, Tianqi
    He, Huagui
    Liang, Xiaojin
    Chen, Fei
    Chen, Zhaoxia
    Liu, Yang
    FRONTIERS IN ENVIRONMENTAL SCIENCE, 2022, 10
  • [22] Exploring Strategies for Training Deep Neural Networks
    Larochelle, Hugo
    Bengio, Yoshua
    Louradour, Jerome
    Lamblin, Pascal
    JOURNAL OF MACHINE LEARNING RESEARCH, 2009, 10 : 1 - 40
  • [23] Exploring strategies for training deep neural networks
    Larochelle, Hugo
    Bengio, Yoshua
    Louradour, Jérôme
    Lamblin, Pascal
    Journal of Machine Learning Research, 2009, 10 : 1 - 40
  • [24] COMPLEX INPUT CONVOLUTIONAL NEURAL NETWORKS FOR WIDE ANGLE SAR ATR
    Wilmanski, Michael
    Kreucher, Chris
    Hero, Alfred
    2016 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2016, : 1037 - 1041
  • [25] Multi-Input Convolutional Neural Networks for Automatic Pollen Classification
    Boldeanu, Mihai
    Cucu, Horia
    Burileanu, Corneliu
    Marmureanu, Luminita
    APPLIED SCIENCES-BASEL, 2021, 11 (24):
  • [26] MULTIPLE POINTS INPUT FOR CONVOLUTIONAL NEURAL NETWORKS IN REPLAY ATTACK DETECTION
    Yoon, Sung-Hyun
    Yu, Ha-Jin
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6444 - 6448
  • [27] SOURCE-SEPARATED AUDIO INPUT FOR ACCELERATING CONVOLUTIONAL NEURAL NETWORKS
    Dominguez, Miguel
    Daigneau, Madeleine
    Ptucha, Raymond
    2017 IEEE WESTERN NEW YORK IMAGE AND SIGNAL PROCESSING WORKSHOP (WNYISPW), 2017,
  • [28] AdaFlow: Aggressive Convolutional Neural Networks Approximation by Leveraging the Input Variability
    Lu, Wenyan
    Yan, Guihai
    Li, Xiaowei
    JOURNAL OF LOW POWER ELECTRONICS, 2018, 14 (04) : 481 - 495
  • [29] Training Convolutional Neural Networks with Limited Training Data for Ear Recognition in the Wild
    Emersic, Ziga
    Stepec, Dejan
    Struc, Vitomir
    Peer, Peter
    2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, : 987 - 994
  • [30] Implementation Challenges and Strategies for Hebbian Learning in Convolutional Neural Networks
    A. V. Demidovskij
    M. S. Kazyulina
    I. G. Salnikov
    A. M. Tugaryov
    A. I. Trutnev
    S. V. Pavlov
    Optical Memory and Neural Networks, 2023, 32 : S252 - S264