Efficient Hardware Accelerator for Compressed Sparse Deep Neural Network

被引:3
|
作者
Xiao, Hao [1 ]
Zhao, Kaikai [1 ]
Liu, Guangzhu [1 ]
机构
[1] HeFei Univ Technol, Sch Microelect, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
deep neural networks; filed programmable gate array; run-length compression; sparse data;
D O I
10.1587/transinf.2020EDL8153
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This work presents a DNN accelerator architecture specifically designed for performing efficient inference on compressed and sparse DNN models. Leveraging the data sparsity, a runtime processing scheme is proposed to deal with the encoded weights and activations directly in the compressed domain without decompressing. Furthermore, a new data flow is proposed to facilitate the reusage of input activations across the fully-connected (FC) layers. The proposed design is implemented and verified using the Xilinx Virtex-7 FPGA. Experimental results show it achieves 1.99x, 1.95x faster and 20.38x, 3.04x more energy efficient than CPU and mGPU platforms, respectively, running AlexNet.
引用
收藏
页码:772 / 775
页数:4
相关论文
共 50 条
  • [1] An Efficient Hardware Accelerator for Sparse Transformer Neural Networks
    Fang, Chao
    Guo, Shouliang
    Wu, Wei
    Lin, Jun
    Wang, Zhongfeng
    Hsu, Ming Kai
    Liu, Lingzhi
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 2670 - 2674
  • [2] A Computing Efficient Hardware Architecture for Sparse Deep Neural Network Computing
    Zhang, Yanwen
    Ouyang, Peng
    Yin, Shouyi
    Zhang, Youguang
    Zhao, Weisheng
    Wei, Shaojun
    2018 14TH IEEE INTERNATIONAL CONFERENCE ON SOLID-STATE AND INTEGRATED CIRCUIT TECHNOLOGY (ICSICT), 2018, : 1261 - 1263
  • [3] An Efficient Hardware Accelerator for Sparse Convolutional Neural Networks on FPGAs
    Lu, Liqiang
    Xie, Jiaming
    Huang, Ruirui
    Zhang, Jiansong
    Lin, Wei
    Liang, Yun
    2019 27TH IEEE ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM), 2019, : 17 - 25
  • [4] Sparse Compressed Spiking Neural Network Accelerator for Object Detection
    Lien, Hong-Han
    Chang, Tian-Sheuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (05) : 2060 - 2069
  • [5] DEEPEYE: A Deeply Tensor-Compressed Neural Network Hardware Accelerator
    Cheng, Yuan
    Li, Guangya
    Wong, Ngai
    Chen, Hai-Bao
    Yu, Hao
    2019 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD), 2019,
  • [6] An Efficient Accelerator Unit for Sparse Convolutional Neural Network
    Zhao, Yulin
    Wang, Donghui
    Wang, Leiou
    TENTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2018), 2018, 10806
  • [7] An Efficient Hardware Accelerator for Block Sparse Convolutional Neural Networks on FPGA
    Yin, Xiaodi
    Wu, Zhipeng
    Li, Dejian
    Shen, Chongfei
    Liu, Yu
    IEEE EMBEDDED SYSTEMS LETTERS, 2024, 16 (02) : 158 - 161
  • [8] An Efficient Hardware Accelerator for Structured Sparse Convolutional Neural Networks on FPGAs
    Zhu, Chaoyang
    Huang, Kejie
    Yang, Shuyuan
    Zhu, Ziqi
    Zhang, Hejia
    Shen, Haibin
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (09) : 1953 - 1965
  • [9] RECOM: An Efficient Resistive Accelerator for Compressed Deep Neural Networks
    Ji, Houxiang
    Song, Linghao
    Jiang, Li
    Li, Ha
    Chen, Yiran
    PROCEEDINGS OF THE 2018 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2018, : 237 - 240
  • [10] An Energy-Efficient Accelerator with Relative-Indexing Memory for Sparse Compressed Convolutional Neural Network
    Wu, I-Chen
    Huang, Po-Tsang
    Lo, Chin-Yang
    Hwang, Wei
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 42 - 45