Efficient Hardware Acceleration of Convolutional Neural Networks

被引:0
|
作者
Kala, S. [1 ]
Jose, Babita R. [1 ]
Mathew, Jimson [2 ]
Nalesh, S. [3 ]
机构
[1] Cochin Univ Sci & Technol, Sch Engn, Kochi, Kerala, India
[2] Indian Inst Technol Patna, Dept Comp Sci & Engn, Patna, Bihar, India
[3] Cochin Univ Sci & Technol, Dept Elect, Cochin, Kerala, India
关键词
Convolutional neural networks; FPGA; high performance; Winograd algorithm;
D O I
10.1109/SOCC46988.2019.1570573948
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional neural networks (CNNs) have emerged as the most efficient technique for solving a host of machine learning tasks. Compute and memory intensive nature of CNN has stimulated lot of work in hardware acceleration of these network models. FPGAs have emerged as a promising approach for accelerating CNNs, due to its high performance, flexibility and energy efficiency. We propose a unified architecture named UniWiG, where both Winograd based convolution and general matrix multiplication (GEMM) can be accelerated using the same set of processing elements. Proposed architecture has been used to accelerate AlexNet and VGG-16 models on FPGA with a performance of 433.63 GOPS and 407.23 GOPS respectively. We have also analyzed the performance with varying Winograd tile sizes and found out the most appropriate tile sizes for maximizing the performance while reducing on-chip memory resource.
引用
收藏
页码:191 / 192
页数:2
相关论文
共 50 条
  • [31] VIBNN: Hardware Acceleration of Bayesian Neural Networks
    Cai, Ruizhe
    Ren, Ao
    Liu, Ning
    Ding, Caiwen
    Wang, Luhao
    Qian, Xuehai
    Pedram, Massoud
    Wang, Yanzhi
    ACM SIGPLAN NOTICES, 2018, 53 (02) : 476 - 488
  • [32] Hardware Design Automation of Convolutional Neural Networks
    Solazzo, Andrea
    Del Sozzo, Emanuele
    De Rose, Irene
    De Silvestri, Matteo
    Durelli, Gianluca C.
    Santambrogio, Marco D.
    2016 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2016, : 224 - 229
  • [33] FPGA Implementation and Acceleration of Convolutional Neural Networks
    Pisharody, Jayanth N.
    Pranav, K. B.
    Ranjitha, M.
    Rajeshwari, B.
    2021 6TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2021,
  • [34] Optimization and acceleration of convolutional neural networks: A survey
    Habib, Gousia
    Qureshi, Shaima
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (07) : 4244 - 4268
  • [35] SLIT: An Energy-Efficient Reconfigurable Hardware Architecture for Deep Convolutional Neural Networks
    Tran, Thi Diem
    Nakashima, Yasuhiko
    IEICE TRANSACTIONS ON ELECTRONICS, 2021, E104C (07) : 319 - 329
  • [36] Albireo: Energy-Efficient Acceleration of Convolutional Neural Networks via Silicon Photonics
    Shiflett, Kyle
    Karanth, Avinash
    Bunescu, Razvan
    Louri, Ahmed
    2021 ACM/IEEE 48TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA 2021), 2021, : 860 - 873
  • [37] Low-Cost IoT Surveillance System Using Hardware-Acceleration and Convolutional Neural Networks
    Lage, Epaminondas S.
    Santos, Rodolfo L.
    Junior, Sandro M. T.
    Andreotti, Fernando
    2019 IEEE 5TH WORLD FORUM ON INTERNET OF THINGS (WF-IOT), 2019, : 931 - 936
  • [38] Digital Hardware Acceleration for Neural Networks: Implementation Considerations
    Petrot, Frederic
    PROCEEDINGS OF THE 2023 34TH INTERNATIONAL WORKSHOP ON RAPID SYSTEM PROTOTYPING, RSP 2023, 2023,
  • [39] Efficient Hardware Architectures for Deep Convolutional Neural Network
    Wang, Jichen
    Lin, Jun
    Wang, Zhongfeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (06) : 1941 - 1953
  • [40] Data and Hardware Efficient Design for Convolutional Neural Network
    Lin, Yue-Jin
    Chang, Tian Sheuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (05) : 1642 - 1651