Hardware Architecture Exploration for Deep Neural Networks

被引:2
|
作者
Zheng, Wenqi [1 ]
Zhao, Yangyi [1 ]
Chen, Yunfan [1 ]
Park, Jinhong [2 ]
Shin, Hyunchul [1 ]
机构
[1] Hanyang Univ, Dept Elect Engn, Ansan, South Korea
[2] Samsung Elect Inc, Suwon, South Korea
关键词
AI architecture; Neural network architecture; CNN; Design space exploration;
D O I
10.1007/s13369-021-05455-4
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Owing to good performance, deep Convolution Neural Networks (CNNs) are rapidly rising in popularity across a broad range of applications. Since high accuracy CNNs are both computation intensive and memory intensive, many researchers have shown significant interest in the accelerator design. Furthermore, the AI chip market size grows and the competition on the performance, cost, and power consumption of the artificial intelligence SoC designs is increasing. Therefore, it is important to develop design techniques and platforms that are useful for the efficient design of optimized AI architectures to satisfy the given specifications in a short design time. In this research, we have developed design space exploration techniques and environments for the optimal design of the overall system including computing modules and memories. Our current design platform is built using NVIDIA Deep Learning Accelerator as a computing model, SRAM as a buffer, and DRAM with GDDR6 as an off-chip memory. We also developed a program to estimate the processing time of a given neural network. By modifying both the on-chip SRAM size and the computing module size, a designer can explore the design space efficiently, and then choose the optimal architecture which shows the minimal cost while satisfying the performance specification. To illustrate the operation of the design platform, two well-known deep CNNs are used, which are YOLOv3 and faster RCNN. This technology can be used to explore and to optimize the hardware architectures of the CNNs so that the cost can be minimized.
引用
收藏
页码:9703 / 9712
页数:10
相关论文
共 50 条
  • [41] A Hardware Accelerator Based on Quantized Weights for Deep Neural Networks
    Sreehari, R.
    Deepu, Vijayasenan
    Arulalan, M. R.
    EMERGING RESEARCH IN ELECTRONICS, COMPUTER SCIENCE AND TECHNOLOGY, ICERECT 2018, 2019, 545 : 1079 - 1091
  • [42] SafeTPU: A Verifiably Secure Hardware Accelerator for Deep Neural Networks
    Collantes, Maria I. Mera
    Ghodsi, Zahra
    Garg, Siddharth
    2020 IEEE 38TH VLSI TEST SYMPOSIUM (VTS 2020), 2020,
  • [43] A Survey of Deep Neural Networks: Deployment Location and Underlying Hardware
    Kotlar, Milos
    Bojic, Dragan
    Punt, Marija
    Milutinovic, Veljko
    2018 14TH SYMPOSIUM ON NEURAL NETWORKS AND APPLICATIONS (NEUREL), 2018,
  • [44] Interpretability of deep neural networks: A review of methods, classification and hardware
    Antamis, Thanasis
    Drosou, Anastasis
    Vafeiadis, Thanasis
    Nizamis, Alexandros
    Ioannidis, Dimosthenis
    Tzovaras, Dimitrios
    NEUROCOMPUTING, 2024, 601
  • [45] Hardware for Quantized Mixed-Precision Deep Neural Networks
    Rios, Andres
    Nava, Patricia
    PROCEEDINGS OF THE 2022 15TH IEEE DALLAS CIRCUITS AND SYSTEMS CONFERENCE (DCAS 2022), 2022,
  • [46] Hardware Efficient Convolution Processing Unit for Deep Neural Networks
    Hazarika, Anakhi
    Poddar, Soumyajit
    Rahaman, Hafizur
    2019 2ND INTERNATIONAL SYMPOSIUM ON DEVICES, CIRCUITS AND SYSTEMS (ISDCS 2019), 2019,
  • [47] An In-depth Comparison of Compilers for Deep Neural Networks on Hardware
    Xing, Yu
    Weng, Jian
    Wang, Yushun
    Sui, Lingzhi
    Shan, Yi
    Wang, Yu
    2019 IEEE INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE AND SYSTEMS (ICESS), 2019,
  • [48] NASA: Neural Architecture Search and Acceleration for Hardware Inspired Hybrid Networks
    Shi, Huihong
    You, Haoran
    Zhao, Yang
    Wang, Zhongfeng
    Lin, Yingyan
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [49] VHDL Synthesizable Hardware Architecture Design of Back Propagation Neural Networks
    Rajeswaran, N.
    Madhu, T.
    Suryakalavathi, M.
    2013 IEEE CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGIES (ICT 2013), 2013, : 445 - 450
  • [50] Configurable N-fold Hardware Architecture for Convolutional Neural Networks
    Baptista, Dario
    Morgado-Dias, F.
    Sousa, Leonel
    2018 INTERNATIONAL CONFERENCE ON BIOMEDICAL ENGINEERING AND APPLICATIONS (ICBEA), 2018, : 92 - 99