Pico-Programmable Neurons to Reduce Computations for Deep Neural Network Accelerators

被引:0
|
作者
Nahvy, Alireza [1 ]
Navabi, Zainalabedin [1 ]
机构
[1] Univ Tehran, Coll Engn, Sch Elect & Comp Engn, Tehran 1439957131, Iran
关键词
Computational reuse; deep neural networks (DNN); multiplication and accumulation (MAC) operation; microprogramed architecture; EFFICIENT; SPARSE; ARCHITECTURE;
D O I
10.1109/TVLSI.2024.3386698
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNNs) have shown impressive success in various fields. As a response to the ever-growing precision demand of DNN applications, more complex computational models are created. The growing computational volume has become a challenge for the power and performance efficiency of DNN accelerators. This article presents a new neural architecture to prevent ineffective and redundant computations by using neurons with memory that have decision-making power. In addition, another local memory is used to keep calculation history for removing redundancy by computational reuse. Sparse computing, as another feature, is supported to remove computations of not only zero weights but also zero bits of each weight. The results on conventional datasets such as IMAGENET show a computational reduction of more than 18 $\times$ -150 $\times$ . This scalable architecture enables 124 GOPS by using 197-mW power.
引用
收藏
页码:1216 / 1227
页数:12
相关论文
共 50 条
  • [1] Review of ASIC accelerators for deep neural network
    Machupalli, Raju
    Hossain, Masum
    Mandal, Mrinal
    MICROPROCESSORS AND MICROSYSTEMS, 2022, 89
  • [2] Approximate Adders for Deep Neural Network Accelerators
    Raghuram, S.
    Shashank, N.
    2022 35TH INTERNATIONAL CONFERENCE ON VLSI DESIGN (VLSID 2022) HELD CONCURRENTLY WITH 2022 21ST INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS (ES 2022), 2022, : 210 - 215
  • [3] A Survey on Memory Subsystems for Deep Neural Network Accelerators
    Asad, Arghavan
    Kaur, Rupinder
    Mohammadi, Farah
    FUTURE INTERNET, 2022, 14 (05):
  • [4] Dynamic Precision Multiplier For Deep Neural Network Accelerators
    Ding, Chen
    Yuxiang, Huan
    Zheng, Lirong
    Zou, Zhuo
    2020 IEEE 33RD INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE (SOCC), 2020, : 180 - 184
  • [5] An Overview of Efficient Interconnection Networks for Deep Neural Network Accelerators
    Nabavinejad, Seyed Morteza
    Baharloo, Mohammad
    Chen, Kun-Chih
    Palesi, Maurizio
    Kogel, Tim
    Ebrahimi, Masoumeh
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2020, 10 (03) : 268 - 282
  • [6] BenQ: Benchmarking Automated Quantization on Deep Neural Network Accelerators
    Wei, Zheng
    Zhang, Xingjun
    Li, Jingbo
    Ji, Zeyu
    Wei, Jia
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 1479 - 1484
  • [7] SEALing Neural Network Models in Encrypted Deep Learning Accelerators
    Zuo, Pengfei
    Hua, Yu
    Liang, Ling
    Xie, Xinfeng
    Hu, Xing
    Xie, Yuan
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 1255 - 1260
  • [8] Tango: A Deep Neural Network Benchmark Suite for Various Accelerators
    Karki, Aajna
    Keshava, Chethan Palangotu
    Shivakumar, Spoorthi Mysore
    Skow, Joshua
    Hegde, Goutam Madhukeshwar
    Jeon, Hyeran
    2019 IEEE INTERNATIONAL SYMPOSIUM ON PERFORMANCE ANALYSIS OF SYSTEMS AND SOFTWARE (ISPASS), 2019, : 137 - 138
  • [9] Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
    Armeniakos, Giorgos
    Zervakis, Georgios
    Soudris, Dimitrios
    Henkel, Joerg
    ACM COMPUTING SURVEYS, 2023, 55 (04)
  • [10] LAMBDA: An Open Framework for Deep Neural Network Accelerators Simulation
    Russo, Enrico
    Palesi, Maurizio
    Monteleone, Salvatore
    Patti, Davide
    Ascia, Giuseppe
    Catania, Vincenzo
    2021 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS AND OTHER AFFILIATED EVENTS (PERCOM WORKSHOPS), 2021, : 161 - 166