Dual-Precision Acceleration of Convolutional Neural Network Computation with Mixed Input and Output Data Reuse

被引:0
|
作者
Hsiao, Shen-Fu [1 ]
Wu, Pei-Hsuan [1 ]
Chen, Jien-Min [1 ]
Chen, Kun-Chih [1 ]
机构
[1] Natl Sun Yat Sen Univ, Dept Comp Sci & Engn, Kaohsiung, Taiwan
关键词
Convolutional Neural Network; Deep Neural Network; Machine Learning; data reuse; hardware accelerator;
D O I
10.1109/iscas.2019.8702094
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Memory access dominates power consumption in hardware acceleration of deep neural networks (DNN) computation due to the movement of huge data and weights. This paper design a DNN accelerator using mixed input and output data reuse scheme to achieve balance between internal memory size and memory access amount, two contradictory design goals in resource limited embedded systems. First, analytical forms for memory size and accesses are derived for different data reuse methods in DNN convolution. After comparing the analysis results across different convolutional layers of the VGG-16 model with different levels of hardware parallelism, we implement a low-cost DNN hardware accelerator using mixed input and output data reuse scheme with 32 processing elements (PEs) operating in parallel. Furthermore, the design supports two precision modes (8-bit and 16-bit) allowing variable precision requirements across DNN layers, resulting in more efficient computation compared with single-precision designs through sharing of hardware resource.
引用
收藏
页数:4
相关论文
共 50 条
  • [1] Dual-Precision Deep Neural Network
    Park, Jae Hyun
    Choi, Ji Sub
    Ko, Jong Hwan
    AIPR 2020: 2020 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND PATTERN RECOGNITION, 2020, : 30 - 34
  • [2] Energy-efficient acceleration of convolutional neural networks using computation reuse
    Ghanbari, Azam
    Modarressi, Mehdi
    JOURNAL OF SYSTEMS ARCHITECTURE, 2022, 126
  • [3] Maximum output discrepancy computation for convolutional neural network compression
    Mo, Zihao
    Xiang, Weiming
    INFORMATION SCIENCES, 2024, 665
  • [4] Event identification of φ-OTDR by dual input deep convolutional neural network
    Feng, Yuhang
    Chen, Cong
    Zhou, Jianjun
    Wen, Mingxu
    Qin, Zujun
    AOPC 2023:OPTIC FIBER GYRO, 2023, 12968
  • [5] Convolutional Neural Network with Discriminant Criterion for Input of Each Neuron in Output Layer
    Ide, Hidenori
    Kurita, Takio
    NEURAL INFORMATION PROCESSING (ICONIP 2018), PT I, 2018, 11301 : 332 - 339
  • [6] A Mixed-Pruning Based Framework for Embedded Convolutional Neural Network Acceleration
    Chang, Xuepeng
    Pan, Huihui
    Lin, Weiyang
    Gao, Huijun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (04) : 1706 - 1715
  • [7] A Speed Regression Using Acceleration Data in a Deep Convolutional Neural Network
    Wang, Karl L.
    Xu, Jingya
    IEEE ACCESS, 2019, 7 : 9351 - 9356
  • [8] INFLUENCE OF INPUT IMAGE CONFIGURATIONS ON OUTPUT OF A CONVOLUTIONAL NEURAL NETWORK TO DETECT CEREBRAL ANEURYSMS
    Watanabe, Kazuhiro
    Anzai, Hitomi
    Juchler, Norman
    Hirsch, Sven
    Bijlenga, Philippe
    Ohta, Makoto
    PROCEEDINGS OF THE ASME INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION, 2019, VOL 3, 2020,
  • [9] Automated Adaptation of Input and Output Data for a Weightless Artificial Neural Network
    McElroy, Ben
    Howells, Gareth
    INTERNATIONAL JOURNAL OF GRID AND DISTRIBUTED COMPUTING, 2011, 4 (03): : 37 - 46
  • [10] Automated adaptation of input and output data for a weightless artificial neural network
    McElroy, Ben
    Howells, Gareth
    International Journal of Database Theory and Application, 2011, 4 (03): : 49 - 58