A Multistage Dataflow Implementation of a Deep Convolutional Neural Network Based on FPGA For High-Speed Object Recognition

被引:0
|
作者
Li, Ning [1 ]
Takaki, Shunpei [1 ]
Tomioka, Yoichi [2 ]
Kitazawa, Hitoshi [1 ]
机构
[1] Tokyo Univ Agr & Technol, 2-24-16 Naka Cho, Koganei, Tokyo, Japan
[2] Univ Aizu Aizu Wakamatsu, Fukushima, Japan
关键词
FPGA Accelerator; Convolutional Neural Network; Image Recognition;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNNs) have progressed significantly in recent years. Novel DNN methods allow tasks such as image and speech recognition to be conducted easily and efficiently, com pared with previous methods that needed to search for valid feature values or algorithms. However, DNN computations typically consume a significant amount of time and high-performance computing resources. To facilitate high-speed object recognition, this article introduces a Deep Convolutional Neural Network (DCNN) accelerator based on a field-programmable gate array (FPGA). Our hardware takes full advantage of the characteristics of convolutional calculation; this allowed us to implement all DCNN layers, from image input to classification, in a single chip. In particular, the dateflow from input to classification is uninterrupted and paralleled. As a result, our implementation achieved a speed of 409.62 giga-operations per second (GOPS), which is approximately twice as fast as the latest reported result. Furthermore, we used the same architecture to implement a Recurrent Convolutional Neural Network (RCNN), which can, in theory, provide better recognition accuracy.
引用
收藏
页码:165 / 168
页数:4
相关论文
共 50 条
  • [31] Object Recognition Algorithm Based on an Improved Convolutional Neural Network
    Zheyi Fan
    Yu Song
    Wei Li
    JournalofBeijingInstituteofTechnology, 2020, 29 (02) : 139 - 145
  • [32] Object Recognition Algorithm Based on an Improved Convolutional Neural Network
    Fan Z.
    Song Y.
    Li W.
    Fan, Zheyi (funye@bit.edu.cn), 1600, Beijing Institute of Technology (29): : 139 - 145
  • [33] High-Speed Tiny Tennis Ball Detection Based on Deep Convolutional Neural Networks
    Tian, Binren
    Zhang, Debing
    Zhang, Chun
    2020 IEEE 14TH INTERNATIONAL CONFERENCE ON ANTI-COUNTERFEITING, SECURITY, AND IDENTIFICATION (ASID), 2020, : 31 - 34
  • [34] High-Speed Object Recognition Based on a Neuromorphic System
    Yang, Zonglin
    Yang, Liren
    Bao, Wendi
    Tao, Liying
    Zeng, Yinuo
    Hu, Die
    Xiong, Jianping
    Shang, Delong
    ELECTRONICS, 2022, 11 (24)
  • [35] Optimizing Loop Operation and Dataflow in FPGA Acceleration of Deep Convolutional Neural Networks
    Ma, Yufei
    Cao, Yu
    Vrudhula, Sarma
    Seo, Jae-sun
    FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2017, : 45 - 54
  • [36] On the Reliability of Convolutional Neural Network Implementation on SRAM-based FPGA
    Du, Boyang
    Azimi, Sarah
    De Sio, Corrado
    Bozzoli, Ludovica
    Sterpone, Luca
    2019 IEEE INTERNATIONAL SYMPOSIUM ON DEFECT AND FAULT TOLERANCE IN VLSI AND NANOTECHNOLOGY SYSTEMS (DFT), 2019,
  • [37] Modulation recognition using an FPGA-based convolutional neural network
    Liu, Xueyuan
    Shang, Jing
    Leong, Philip H. W.
    Liu, Cheng
    2019 22ND INTERNATIONAL CONFERENCE ON ELECTRICAL MACHINES AND SYSTEMS (ICEMS 2019), 2019, : 3165 - 3170
  • [38] Recurrent Convolutional Neural Network for Object Recognition
    Liang, Ming
    Hu, Xiaolin
    2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, : 3367 - 3375
  • [39] Design and implementation of an FPGA architecture for high-speed network feature extraction
    Pati, Sailesh
    Narayanan, Ramanathan
    Memik, Gokhan
    Choudhary, Alok
    Zambreno, Joseph
    ICFPT 2007: INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY, PROCEEDINGS, 2007, : 49 - +
  • [40] Memristor Crossbar Deep Network Implementation Based on a Convolutional Neural Network
    Yakopcic, Chris
    Alom, Md Zahangir
    Taha, Tarek M.
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 963 - 970