Hardware Architecture Exploration for Deep Neural Networks

被引:2
|
作者
Zheng, Wenqi [1 ]
Zhao, Yangyi [1 ]
Chen, Yunfan [1 ]
Park, Jinhong [2 ]
Shin, Hyunchul [1 ]
机构
[1] Hanyang Univ, Dept Elect Engn, Ansan, South Korea
[2] Samsung Elect Inc, Suwon, South Korea
关键词
AI architecture; Neural network architecture; CNN; Design space exploration;
D O I
10.1007/s13369-021-05455-4
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Owing to good performance, deep Convolution Neural Networks (CNNs) are rapidly rising in popularity across a broad range of applications. Since high accuracy CNNs are both computation intensive and memory intensive, many researchers have shown significant interest in the accelerator design. Furthermore, the AI chip market size grows and the competition on the performance, cost, and power consumption of the artificial intelligence SoC designs is increasing. Therefore, it is important to develop design techniques and platforms that are useful for the efficient design of optimized AI architectures to satisfy the given specifications in a short design time. In this research, we have developed design space exploration techniques and environments for the optimal design of the overall system including computing modules and memories. Our current design platform is built using NVIDIA Deep Learning Accelerator as a computing model, SRAM as a buffer, and DRAM with GDDR6 as an off-chip memory. We also developed a program to estimate the processing time of a given neural network. By modifying both the on-chip SRAM size and the computing module size, a designer can explore the design space efficiently, and then choose the optimal architecture which shows the minimal cost while satisfying the performance specification. To illustrate the operation of the design platform, two well-known deep CNNs are used, which are YOLOv3 and faster RCNN. This technology can be used to explore and to optimize the hardware architectures of the CNNs so that the cost can be minimized.
引用
收藏
页码:9703 / 9712
页数:10
相关论文
共 50 条
  • [21] A POPULATION CODING HARDWARE ARCHITECTURE FOR SPIKING NEURAL NETWORKS APPLICATIONS
    Nuno-Maganda, Marco
    Arias-Estrada, Miguel
    Torres Huitzil, Cesar
    Girau, Bernard
    2009 5TH SOUTHERN CONFERENCE ON PROGRAMMABLE LOGIC, PROCEEDINGS, 2009, : 83 - +
  • [22] A Hardware Architecture for Image Clustering Using Spiking Neural Networks
    Aurelio Nuno-Maganda, Marco
    Arias-Estrada, Miguel
    Torres-Huitzil, Cesar
    Hugo Aviles-Arriaga, Hector
    Hernandez-Mier, Yahir
    Morales-Sandoval, Miguel
    2012 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2012, : 261 - 266
  • [23] Layered tile architecture for efficient hardware spiking neural networks
    Wan, Lei
    Liu, Junxiu
    Harkin, Jim
    McDaid, Liam
    Luo, Yuling
    MICROPROCESSORS AND MICROSYSTEMS, 2017, 53 : 21 - 32
  • [24] Low Cost Interconnected Architecture for the Hardware Spiking Neural Networks
    Luo, Yuling
    Wan, Lei
    Liu, Junxiu
    Harkin, Jim
    McDaid, Liam
    Cao, Yi
    Ding, Xuemei
    FRONTIERS IN NEUROSCIENCE, 2018, 12
  • [25] Novel architecture and synapse design for hardware implementations of neural networks
    McGinnity, TM
    Roche, B
    Maguire, LP
    McDaid, LJ
    COMPUTERS & ELECTRICAL ENGINEERING, 1998, 24 (1-2) : 75 - 87
  • [26] Novel architecture and synapse design for hardware implementations of neural networks
    Faculty of Engineering, Magee College, University of Ulster, Northland Road, Derry, BT48 7JL, United Kingdom
    Comput Electr Eng, 1-2 (75-87):
  • [28] The Impact of Architecture on the Deep Neural Networks Training
    Rozycki, Pawel
    Kolbusz, Janusz
    Malinowski, Aleksander
    Wilamowski, Bogdan
    2019 12TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2019, : 41 - 46
  • [29] LiteLSTM Architecture for Deep Recurrent Neural Networks
    Elsayed, Nelly
    ElSayed, Zag
    Maida, Anthony S.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1304 - 1308
  • [30] An Architecture to Accelerate Convolution in Deep Neural Networks
    Ardakani, Arash
    Condo, Carlo
    Ahmadi, Mehdi
    Gross, Warren J.
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (04) : 1349 - 1362