A Flexible Sparsity-Aware Accelerator with High Sensitivity and Efficient Operation for Convolutional Neural Networks

被引:1
|
作者
Yuan, Haiying [1 ]
Zeng, Zhiyong [1 ]
Cheng, Junpeng [1 ]
Li, Minghao [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Sparsity perceptron; Parallel computing; FPGA accelerator;
D O I
10.1007/s00034-022-01992-x
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In view of the technical challenge that convolutional neural networks involve in a large amount of computation caused by the information redundancy of the interlayer activation, a flexible sparsity-aware accelerator is proposed in this paper. It realizes the basic data transmission with coarse-grained control and realizes the transmission of sparse data with fine-grained control. In addition, the corresponding data arrangement scheme is designed to fully utilize the off-chip bandwidth. In order to improve the inference performance without accuracy reduction, the sparse activation is compressed to eliminate ineffectual activation while preserving topology information with the sparsity perceptron module. To improve power efficiency, the computational load is rationally allocated for multiplication accumulator array, and the convolution operation is decoupled by adder tree with FIFO. The accelerator is implemented on Xilinx VCU108, and 97.27% of the operations are non-zero activation operations. The accelerator running in sparsity mode is more than 2.5 times faster than that in density mode, and power consumption is reduced to 8.3 W. Furthermore, this flexible sparsity-aware accelerator architecture can be widely applied to large-scale deep convolutional neural networks.
引用
收藏
页码:4370 / 4389
页数:20
相关论文
共 50 条
  • [1] A Flexible Sparsity-Aware Accelerator with High Sensitivity and Efficient Operation for Convolutional Neural Networks
    Haiying Yuan
    Zhiyong Zeng
    Junpeng Cheng
    Minghao Li
    Circuits, Systems, and Signal Processing, 2022, 41 : 4370 - 4389
  • [2] A Sparsity-Aware Convolutional Neural Network Accelerator with Flexible Parallelism
    Yuan H.-Y.
    Zeng Z.-Y.
    Cheng J.-P.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2022, 50 (08): : 1811 - 1818
  • [3] SATA: Sparsity-Aware Training Accelerator for Spiking Neural Networks
    Yin, Ruokai
    Moitra, Abhishek
    Bhattacharjee, Abhiroop
    Kim, Youngeun
    Panda, Priyadarshini
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (06) : 1926 - 1938
  • [4] A Convolutional Spiking Neural Network Accelerator with the Sparsity-aware Memory and Compressed Weights
    Liu, Hanqing
    Cui, Xiaole
    Zhang, Sunrui
    Yin, Mingqi
    Jiang, Yuanyuan
    Cui, Xiaoxin
    2024 IEEE 35TH INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS, ASAP 2024, 2024, : 163 - 171
  • [5] Quantization and sparsity-aware processing for energy-efficient NVM-based convolutional neural networks
    Bao, Han
    Qin, Yifan
    Chen, Jia
    Yang, Ling
    Li, Jiancong
    Zhou, Houji
    Li, Yi
    Miao, Xiangshui
    FRONTIERS IN ELECTRONICS, 2022, 3
  • [6] Sparsity-Aware Caches to Accelerate Deep Neural Networks
    Ganesan, Vinod
    Sen, Sanchari
    Kumar, Pratyush
    Gala, Neel
    Veezhinathan, Kamakoti
    Raghunathan, Anand
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 85 - 90
  • [7] Sparsity-aware generalization theory for deep neural networks
    Muthukumar, Ramchandran
    Sulam, Jeremias
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [8] Sparsity-Aware Orthogonal Initialization of Deep Neural Networks
    Esguerra, Kiara
    Nasir, Muneeb
    Tang, Tong Boon
    Tumian, Afidalina
    Ho, Eric Tatt Wei
    IEEE ACCESS, 2023, 11 : 74165 - 74181
  • [9] TNSS:Two-Nibble Sparsity-Aware Stride Decomposing Acceleration for Convolutional Neural Networks
    Huang, Yun-Yin
    Chen, Yu-Guang
    Jou, Jing-Yang
    2024 IEEE THE 20TH ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS, APCCAS 2024, 2024, : 795 - 799
  • [10] An Efficient and Flexible Accelerator Design for Sparse Convolutional Neural Networks
    Xie, Xiaoru
    Lin, Jun
    Wang, Zhongfeng
    Wei, Jinghe
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (07) : 2936 - 2949