TORRES: A Resource-Efficient Inference Processor for Binary Convolutional Neural Networks Based on Locality-Aware Operation Skipping

被引:3
|
作者
Lee, Su-Jung [1 ]
Kwak, Gil-Ho [1 ]
Kim, Tae-Hwan [1 ]
机构
[1] Korea Aerosp Univ, Sch Elect & Informat Engn, 76 Hanggongdaehak Ro, Goyang Si 10540, South Korea
关键词
binary convolutional neural networks; processor; inference; resource efficiency; field-programmable gate array;
D O I
10.3390/electronics11213534
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A binary convolutional neural network (BCNN) is a neural network promising to realize analysis of visual imagery in low-cost resource-limited devices. This study presents an efficient inference processor for BCNNs, named TORRES. TORRES performs inference efficiently, skipping operations based on the spatial locality inherent in feature maps. The training process is regularized with the objective of skipping more operations. The microarchitecture is designed to skip operations and generate addresses efficiently with low resource usage. A prototype inference system based on TORRES has been implemented in a 28 nm field-programmable gate array, and its functionality has been verified for practical inference tasks. Implemented with 2.31 K LUTs, TORRES achieves the inference speed of 291.2 GOP/s, exhibiting the resource efficiency of 126.06 MOP/s/LUT. The resource efficiency of TORRES is 1.45 times higher than that of the state-of-the-art work.
引用
收藏
页数:16
相关论文
共 20 条
  • [1] Latency and accuracy optimization for binary neural network inference with locality-aware operation skipping
    Lee, S. -J.
    Kim, T. -H.
    ELECTRONICS LETTERS, 2024, 60 (02)
  • [2] A Resource-Efficient Inference Accelerator for Binary Convolutional Neural Networks
    Kim, Tae-Hwan
    Shin, Jihoon
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2021, 68 (01) : 451 - 455
  • [3] Ristretto: A Framework for Empirical Study of Resource-Efficient Inference in Convolutional Neural Networks
    Gysel, Philipp
    Pimentel, Jon
    Motamedi, Mohammad
    Ghiasi, Soheil
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (11) : 5784 - 5789
  • [4] A Resource-Efficient Keyword Spotting System Based on a One-Dimensional Binary Convolutional Neural Network
    Yoon, Jinsung
    Kim, Neungyun
    Lee, Donghyun
    Lee, Su-Jung
    Kwak, Gil-Ho
    Kim, Tae-Hwan
    ELECTRONICS, 2023, 12 (18)
  • [5] A Flexible Sparsity-Aware Accelerator with High Sensitivity and Efficient Operation for Convolutional Neural Networks
    Haiying Yuan
    Zhiyong Zeng
    Junpeng Cheng
    Minghao Li
    Circuits, Systems, and Signal Processing, 2022, 41 : 4370 - 4389
  • [6] A Flexible Sparsity-Aware Accelerator with High Sensitivity and Efficient Operation for Convolutional Neural Networks
    Yuan, Haiying
    Zeng, Zhiyong
    Cheng, Junpeng
    Li, Minghao
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2022, 41 (08) : 4370 - 4389
  • [7] Context-Aware Tree-Based Convolutional Neural Networks for Natural Language Inference
    Meng, Zhao
    Mou, Lili
    Li, Ge
    Jin, Zhi
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2016, 2016, 9983 : 515 - 526
  • [8] Enhancing quantum state tomography via resource-efficient attention-based neural networks
    Palmieri, Adriano Macarone
    Muller-Rigat, Guillem
    Srivastava, Anubhav Kumar
    Lewenstein, Maciej
    Rajchel-Mieldzioc, Grzegorz
    Plodzien, Marcin
    PHYSICAL REVIEW RESEARCH, 2024, 6 (03):
  • [9] ROSETTA: A Resource and Energy-Efficient Inference Processor for Recurrent Neural Networks Based on Programmable Data Formats and Fine Activation Pruning
    Kim, Jiho
    Kim, Tae-Hwan
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2023, 11 (03) : 650 - 663
  • [10] IOTA: a 1.7-TOP/J inference processor for binary convolutional neural networks with 4.7 K LUTs in a tiny FPGA
    Kim, T.
    Shin, J.
    Choi, K.
    ELECTRONICS LETTERS, 2020, 56 (20) : 1041 - 1043