A Flexible Sparsity-Aware Accelerator with High Sensitivity and Efficient Operation for Convolutional Neural Networks

被引:1
|
作者
Yuan, Haiying [1 ]
Zeng, Zhiyong [1 ]
Cheng, Junpeng [1 ]
Li, Minghao [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Sparsity perceptron; Parallel computing; FPGA accelerator;
D O I
10.1007/s00034-022-01992-x
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In view of the technical challenge that convolutional neural networks involve in a large amount of computation caused by the information redundancy of the interlayer activation, a flexible sparsity-aware accelerator is proposed in this paper. It realizes the basic data transmission with coarse-grained control and realizes the transmission of sparse data with fine-grained control. In addition, the corresponding data arrangement scheme is designed to fully utilize the off-chip bandwidth. In order to improve the inference performance without accuracy reduction, the sparse activation is compressed to eliminate ineffectual activation while preserving topology information with the sparsity perceptron module. To improve power efficiency, the computational load is rationally allocated for multiplication accumulator array, and the convolution operation is decoupled by adder tree with FIFO. The accelerator is implemented on Xilinx VCU108, and 97.27% of the operations are non-zero activation operations. The accelerator running in sparsity mode is more than 2.5 times faster than that in density mode, and power consumption is reduced to 8.3 W. Furthermore, this flexible sparsity-aware accelerator architecture can be widely applied to large-scale deep convolutional neural networks.
引用
收藏
页码:4370 / 4389
页数:20
相关论文
共 50 条
  • [41] SuperHCA: An Efficient Deep-Learning Edge Super-Resolution Accelerator With Sparsity-Aware Heterogeneous Core Architecture
    Hu, Zhicheng
    Zeng, Jiahao
    Zhao, Xin
    Zhou, Liang
    Chang, Liang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, : 5420 - 5431
  • [42] TNPU: An Efficient Accelerator Architecture for Training Convolutional Neural Networks
    Li, Jiajun
    Yan, Guihai
    Lu, Wenyan
    Jiang, Shuhao
    Gong, Shijun
    Wu, Jingya
    Yan, Junchao
    Li, Xiaowei
    24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 450 - 455
  • [43] An Efficient Hardware Accelerator for Sparse Convolutional Neural Networks on FPGAs
    Lu, Liqiang
    Xie, Jiaming
    Huang, Ruirui
    Zhang, Jiansong
    Lin, Wei
    Liang, Yun
    2019 27TH IEEE ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM), 2019, : 17 - 25
  • [44] WRA-SS: A High-Performance Accelerator Integrating Winograd With Structured Sparsity for Convolutional Neural Networks
    Yang, Chen
    Meng, Yishuo
    Xi, Jiawei
    Xiang, Siwei
    Wang, Jianfei
    Mei, Kuizhi
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2024, 32 (01) : 164 - 177
  • [45] Sparsity Enables Data and Energy Efficient Spiking Convolutional Neural Networks
    Bhatt, Varun
    Ganguly, Udayan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 : 263 - 272
  • [46] SparseTrain: Exploiting Dataflow Sparsity for Efficient Convolutional Neural Networks Training
    Dai, Pengcheng
    Yang, Jianlei
    Ye, Xucheng
    Cheng, Xingzhou
    Luo, Junyu
    Song, Linghao
    Chen, Yiran
    Zhao, Weisheng
    PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [47] An Energy-Efficient and Flexible Accelerator based on Reconfigurable Computing for Multiple Deep Convolutional Neural Networks
    Yang, Chen
    Zhang, HaiBo
    Wang, XiaoLi
    Geng, Li
    2018 14TH IEEE INTERNATIONAL CONFERENCE ON SOLID-STATE AND INTEGRATED CIRCUIT TECHNOLOGY (ICSICT), 2018, : 1389 - 1391
  • [48] Hardware Flexible Systolic Architecture for Convolution Accelerator in Convolutional Neural Networks
    Aguirre-Alvarez, Paulo Aaron
    Diaz-Carmona, Javier
    Arredondo-Velazquez, Moises
    2022 45TH INTERNATIONAL CONFERENCE ON TELECOMMUNICATIONS AND SIGNAL PROCESSING, TSP, 2022, : 305 - 309
  • [49] Sparsity-aware adaptive link combination approach over distributed networks
    Lu, Songtao
    Nascimento, V. H.
    Sun, Jinping
    Wang, Zhuangji
    ELECTRONICS LETTERS, 2014, 50 (18) : 1285 - U128
  • [50] STICKER: An Energy-Efficient Multi-Sparsity Compatible Accelerator for Convolutional Neural Networks in 65-nm CMOS
    Yuan, Zhe
    Liu, Yongpan
    Yue, Jinshan
    Yang, Yixiong
    Wang, Jingyu
    Feng, Xiaoyu
    Zhao, Jian
    Li, Xueqing
    Yang, Huazhong
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2020, 55 (02) : 465 - 477