A Flexible Sparsity-Aware Accelerator with High Sensitivity and Efficient Operation for Convolutional Neural Networks

被引:1
|
作者
Yuan, Haiying [1 ]
Zeng, Zhiyong [1 ]
Cheng, Junpeng [1 ]
Li, Minghao [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Sparsity perceptron; Parallel computing; FPGA accelerator;
D O I
10.1007/s00034-022-01992-x
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In view of the technical challenge that convolutional neural networks involve in a large amount of computation caused by the information redundancy of the interlayer activation, a flexible sparsity-aware accelerator is proposed in this paper. It realizes the basic data transmission with coarse-grained control and realizes the transmission of sparse data with fine-grained control. In addition, the corresponding data arrangement scheme is designed to fully utilize the off-chip bandwidth. In order to improve the inference performance without accuracy reduction, the sparse activation is compressed to eliminate ineffectual activation while preserving topology information with the sparsity perceptron module. To improve power efficiency, the computational load is rationally allocated for multiplication accumulator array, and the convolution operation is decoupled by adder tree with FIFO. The accelerator is implemented on Xilinx VCU108, and 97.27% of the operations are non-zero activation operations. The accelerator running in sparsity mode is more than 2.5 times faster than that in density mode, and power consumption is reduced to 8.3 W. Furthermore, this flexible sparsity-aware accelerator architecture can be widely applied to large-scale deep convolutional neural networks.
引用
收藏
页码:4370 / 4389
页数:20
相关论文
共 50 条
  • [31] SpeqNets: Sparsity-aware Permutation-equivariant Graph Networks
    Morris, Christopher
    Rattan, Gaurav
    Kiefer, Sandra
    Ravanbkash, Siamak
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [32] Snowflake: An Efficient Hardware Accelerator for Convolutional Neural Networks
    Gokhale, Vinayak
    Zaidy, Aliasger
    Chang, Andre Xian Ming
    Culurciello, Eugenio
    2017 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2017, : 2082 - 2085
  • [33] An Efficient Reconfigurable Hardware Accelerator for Convolutional Neural Networks
    Ansari, Anaam
    Gunnam, Kiran
    Ogunfunmi, Tokunbo
    2017 FIFTY-FIRST ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2017, : 1337 - 1341
  • [34] Sparsity-Aware Deep Learning Accelerator Design Supporting CNN and LSTM Operations
    Hsiao, Shen-Fu
    Chang, Hsuan-Jui
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [35] An Energy-Efficient Unstructured Sparsity-Aware Deep SNN Accelerator With 3-D Computation Array
    Fang, Chaoming
    Shen, Ziyang
    Wang, Zongsheng
    Wang, Chuanqing
    Zhao, Shiqi
    Tian, Fengshi
    Yang, Jie
    Sawan, Mohamad
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2025, 60 (03) : 977 - 989
  • [36] FlexFlow: A Flexible Dataflow Accelerator Architecture for Convolutional Neural Networks
    Lu, Wenyan
    Yan, Guihai
    Li, Jiajun
    Gong, Shijun
    Han, Yinhe
    Li, Xiaowei
    2017 23RD IEEE INTERNATIONAL SYMPOSIUM ON HIGH PERFORMANCE COMPUTER ARCHITECTURE (HPCA), 2017, : 553 - 564
  • [37] Flexible Quantization for Efficient Convolutional Neural Networks
    Zacchigna, Federico Giordano
    Lew, Sergio
    Lutenberg, Ariel
    ELECTRONICS, 2024, 13 (10)
  • [38] RNA: A Flexible and Efficient Accelerator Based on Dynamically Reconfigurable Computing for Multiple Convolutional Neural Networks
    Yang, Chen
    Hou, Jia
    Wang, Yizhou
    Zhang, Haibo
    Wang, Xiaoli
    Geng, Li
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2022, 31 (16)
  • [39] Design of a Sparsity-Aware Reconfigurable Deep Learning Accelerator Supporting Various Types of Operations
    Hsiao, Shen-Fu
    Chen, Kun-Chih
    Lin, Chih-Chien
    Chang, Hsuan-Jui
    Tsai, Bo-Ching
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2020, 10 (03) : 376 - 387
  • [40] NEM-GNN: DAC/ADC-less, Scalable, Reconfigurable, Graph and Sparsity-Aware Near-Memory Accelerator for Graph Neural Networks
    Raman, Siddhartha Raman Sundara
    John, Lizy
    Kulkarni, Jaydeep P.
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2024, 21 (02)