AdaPT: Fast Emulation of Approximate DNN Accelerators in PyTorch

被引:11
|
作者
Danopoulos, Dimitrios [1 ]
Zervakis, Georgios [2 ]
Siozios, Kostas [3 ]
Soudris, Dimitrios [1 ]
Henkel, Joerg [2 ]
机构
[1] Natl Tech Univ Athens, Sch Elect & Comp Engn, Athens 15780, Greece
[2] Karlsruhe Inst Technol, Chair Embedded Syst, D-76131 Karlsruhe, Germany
[3] Aristotle Univ Thessaloniki, Dept Phys, Thessaloniki 54124, Greece
关键词
Accelerator; approximate computing; deep neural network (DNN); PyTorch; quantization; DESIGN;
D O I
10.1109/TCAD.2022.3212645
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Current state-of-the-art employs approximate multipliers to address the highly increased power demands of deep neural network (DNN) accelerators. However, evaluating the accuracy of approximate DNNs is cumbersome due to the lack of adequate support for approximate arithmetic in DNN frameworks. We address this inefficiency by presenting AdaPT, a fast emulation framework that extends PyTorch to support approximate inference as well as approximation-aware retraining. AdaPT can be seamlessly deployed and is compatible with the most DNNs. We evaluate the framework on several DNN models and application fields, including CNNs, LSTMs, and GANs for a number of approximate multipliers with distinct bitwidth values. The results show substantial error recovery from approximate retraining and reduced inference time up to 53.9x with respect to the baseline approximate implementation.
引用
收藏
页码:2074 / 2078
页数:5
相关论文
共 50 条
  • [41] A DNN Protection Solution for PIM accelerators with Model Compression
    Zhao, Lei
    Zhang, Youtao
    Yang, Jun
    2022 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2022), 2022, : 320 - 325
  • [42] Lightning Talk: Efficiency and Programmability of DNN Accelerators and GPUs
    Ro, Won Woo
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [43] Special Session: Approximation and Fault Resiliency of DNN Accelerators
    Ahmadilivani, Mohammad Hasan
    Barbareschi, Mario
    Barone, Salvatore
    Bosio, Alberto
    Daneshtalab, Masoud
    Della Torca, Salvatore
    Gavarini, Gabriele
    Jenihhin, Maksim
    Raik, Jaan
    Ruospo, Annachiara
    Sanchez, Ernesto
    Taheri, Mahdi
    2023 IEEE 41ST VLSI TEST SYMPOSIUM, VTS, 2023,
  • [44] Fault Resilience Techniques for Flash Memory of DNN Accelerators
    Lu, Shyue-Kung
    Wu, Yu-Sheng
    Hong, Jin-Hua
    Miyase, Kohei
    2022 IEEE INTERNATIONAL TEST CONFERENCE IN ASIA (ITC-ASIA 2022), 2022, : 1 - 6
  • [45] NeuroSpector: Systematic Optimization of Dataflow Scheduling in DNN Accelerators
    Park, Chanho
    Kim, Bogil
    Ryu, Sungmin
    Song, William J.
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (08) : 2279 - 2294
  • [46] A survey on modeling and improving reliability of DNN algorithms and accelerators
    Mittal, Sparsh
    JOURNAL OF SYSTEMS ARCHITECTURE, 2020, 104
  • [47] Fault Resilience Techniques for Flash Memory of DNN Accelerators
    Lu, Shyue-Kung
    Wu, Yu-Sheng
    Hong, Jin-Hua
    Miyase, Kohei
    2022 IEEE INTERNATIONAL TEST CONFERENCE (ITC), 2022, : 591 - 600
  • [48] Heterogeneous Dataflow Accelerators for Multi-DNN Workloads
    Kwon, Hyoukjun
    Lai, Liangzhen
    Pellauer, Michael
    Krishna, Tushar
    Chen, Yu-Hsin
    Chandra, Vikas
    2021 27TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE (HPCA 2021), 2021, : 71 - 83
  • [49] Coordinated Batching and DVFS for DNN Inference on GPU Accelerators
    Nabavinejad, Seyed Morteza
    Reda, Sherief
    Ebrahimi, Masoumeh
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (10) : 2496 - 2508
  • [50] Exploring RISC-V Based DNN Accelerators
    Liu, Qiankun
    Amiri, Sam
    Ost, Luciano
    2024 IEEE INTERNATIONAL CONFERENCE ON OMNI-LAYER INTELLIGENT SYSTEMS, COINS 2024, 2024, : 30 - 34