Hardware Efficient Convolution Processing Unit for Deep Neural Networks

被引:2
|
作者
Hazarika, Anakhi [1 ]
Poddar, Soumyajit [1 ]
Rahaman, Hafizur [2 ]
机构
[1] Indian Inst Informat Technol Guwahati, Gauhati 781015, India
[2] Indian Inst Engn Sci & Technol, Sibpur 711103, Howrah, India
关键词
Deep Neural Network; CNN Hardware Accelerator; Field Programmable Gate Array (FPGA);
D O I
10.1109/isdcs.2019.8719278
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Network (CNN) is a type of deep neural networks that are commonly used for object detection and classification. State-of-the-art hardware for training and inference of CNN architectures require a considerable amount of computation and memory intensive resources. CNN achieves greater accuracy at the cost of high computational complexity and large power consumption. To optimize the memory requirement, processing speed and power, it is crucial to design more efficient accelerator architecture for CNN computation. In this work, an overlap of spatially adjacent data is exploited in order to parallelize the movement of data. A fast, re-configurable hardware accelerator architecture along with optimized kernel design suitable for a variety of CNN models is proposed. Our design achieves 2.1x computational benefits over state-of-the-art accelerator architectures.
引用
收藏
页数:4
相关论文
共 50 条
  • [31] Contraction of Dynamically Masked Deep Neural Networks for Efficient Video Processing
    Rueckauer, Bodo
    Liu, Shih-Chii
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (02) : 621 - 633
  • [32] CompAcc: Efficient Hardware Realization for Processing Compressed Neural Networks Using Accumulator Arrays
    Ji, Zexi
    Jung, Wanyeong
    Woo, Jongchan
    Sethi, Khushal
    Lu, Shih-Lien
    Chandrakasan, Anantha P.
    2020 IEEE ASIAN SOLID-STATE CIRCUITS CONFERENCE (A-SSCC), 2020,
  • [33] Hardware Compilation of Deep Neural Networks: An Overview
    Zhao, Ruizhe
    Liu, Shuanglong
    Ng, Ho-Cheung
    Wang, Erwei
    Davis, James J.
    Niu, Xinyu
    Wang, Xiwei
    Shi, Huifeng
    Constantinides, George A.
    Cheung, Peter Y. K.
    Luk, Wayne
    2018 IEEE 29TH INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS (ASAP), 2018, : 120 - 127
  • [34] Hardware Architecture Exploration for Deep Neural Networks
    Zheng, Wenqi
    Zhao, Yangyi
    Chen, Yunfan
    Park, Jinhong
    Shin, Hyunchul
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2021, 46 (10) : 9703 - 9712
  • [35] Formal Verification of Deep Neural Networks in Hardware
    Saji, Sincy Ann
    Agrawal, Shreyansh
    Sood, Surinder
    2022 IEEE WOMEN IN TECHNOLOGY CONFERENCE (WINTECHCON): SMARTER TECHNOLOGIES FOR A SUSTAINABLE AND HYPER-CONNECTED WORLD, 2022,
  • [36] Hardware Architecture Exploration for Deep Neural Networks
    Wenqi Zheng
    Yangyi Zhao
    Yunfan Chen
    Jinhong Park
    Hyunchul Shin
    Arabian Journal for Science and Engineering, 2021, 46 : 9703 - 9712
  • [37] SLIT: An Energy-Efficient Reconfigurable Hardware Architecture for Deep Convolutional Neural Networks
    Tran, Thi Diem
    Nakashima, Yasuhiko
    IEICE TRANSACTIONS ON ELECTRONICS, 2021, E104C (07) : 319 - 329
  • [38] An Efficient Hardware Implementation of Activation Functions Using Stochastic Computing for Deep Neural Networks
    Van-Tinh Nguyen
    Tieu-Khanh Luong
    Han Le Duc
    Van-Phuc Hoang
    2018 IEEE 12TH INTERNATIONAL SYMPOSIUM ON EMBEDDED MULTICORE/MANY-CORE SYSTEMS-ON-CHIP (MCSOC 2018), 2018, : 233 - 236
  • [39] Automated design of error-resilient and hardware-efficient deep neural networks
    Schorn, Christoph
    Elsken, Thomas
    Vogel, Sebastian
    Runge, Armin
    Guntoro, Andre
    Ascheid, Gerd
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (24): : 18327 - 18345
  • [40] Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations
    Boo, Yoonho
    Sung, Wonyong
    2017 IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2017,