LV: Latency-Versatile Floating-Point Engine for High-Performance Deep Neural Networks

被引:1
|
作者
Lo, Yun-Chen [1 ]
Tsai, Yu-Chih [1 ]
Liu, Ren-Shuo [1 ]
机构
[1] Natl Tsing Hua Univ, Dept Elect Engn, Hsinchu 300044, Taiwan
关键词
Index Terms-Approximate computation; floating point; latency-versatile architecture;
D O I
10.1109/LCA.2023.3287096
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Computing latency is an important system metric for Deep Neural Networks (DNNs) accelerators. To reduce latency, this work proposes LV, a latency-versatile floating-point engine (FP-PE), which contains the following key contributions: 1) an approximate bit-versatile multiplier-and-accumulate (BV-MAC) unit with early shifter and 2) an on-demand fixed-point-to-floating-point conversion (FXP2FP) unit. The extensive experimental results show that LV outperforms baseline FP-PE and redundancy-aware FP-PE by up to 2.12x and 1.3x speedup using TSMC 40-nm technology, achieving comparable accuracy on the ImageNet classification tasks.
引用
收藏
页码:125 / 128
页数:4
相关论文
共 50 条
  • [1] HIGH-PERFORMANCE FLOATING-POINT IMPLEMENTATION USING FPGAS
    Parker, Michael
    MILCOM 2009 - 2009 IEEE MILITARY COMMUNICATIONS CONFERENCE, VOLS 1-4, 2009, : 323 - 327
  • [2] GENERATING HIGH-PERFORMANCE CUSTOM FLOATING-POINT PIPELINES
    de Dinechin, Florent
    Klein, Cristian
    Pasca, Bogdan
    FPL: 2009 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS, 2009, : 59 - 64
  • [3] A High-Performance Accelerator for Floating-Point Matrix Multiplication
    Jia, Xun
    Wu, Gunning
    Xie, Xianghui
    2017 15TH IEEE INTERNATIONAL SYMPOSIUM ON PARALLEL AND DISTRIBUTED PROCESSING WITH APPLICATIONS AND 2017 16TH IEEE INTERNATIONAL CONFERENCE ON UBIQUITOUS COMPUTING AND COMMUNICATIONS (ISPA/IUCC 2017), 2017, : 396 - 402
  • [4] Tunable Floating-Point for Artificial Neural Networks
    Franceschi, Marta
    Nannarelli, Alberto
    Valle, Maurizio
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS AND SYSTEMS (ICECS), 2018, : 289 - 292
  • [5] High-Performance Computation in Residue Number System Using Floating-Point Arithmetic
    Isupov, Konstantin
    COMPUTATION, 2021, 9 (02) : 1 - 15
  • [6] HIGH-PERFORMANCE FPGA-BASED FLOATING-POINT ADDER WITH THREE INPUTS
    Guntoro, Andre
    Glesner, Manfred
    2008 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE AND LOGIC APPLICATIONS, VOLS 1 AND 2, 2008, : 626 - 629
  • [7] Floating-point division on programmable high-performance signal-processing hardware
    Pilz, NA
    Adamson, K
    6TH WORLD MULTICONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL IX, PROCEEDINGS: IMAGE, ACOUSTIC, SPEECH AND SIGNAL PROCESSING II, 2002, : 519 - 524
  • [8] Precision Exploration of Floating-Point Arithmetic for Spiking Neural Networks
    Kwak, Myeongjin
    Seo, Hyoju
    Kim, Yongtae
    18TH INTERNATIONAL SOC DESIGN CONFERENCE 2021 (ISOCC 2021), 2021, : 71 - 72
  • [9] High-performance, deep neural networks with sub-microsecond latency on FPGAs for trigger applications
    Nottbeck, Noel
    Schmitt, Christian
    Buescher, Volker
    19TH INTERNATIONAL WORKSHOP ON ADVANCED COMPUTING AND ANALYSIS TECHNIQUES IN PHYSICS RESEARCH, 2020, 1525
  • [10] Open source high performance floating-point modules
    Hemmert, K. Scott
    Underwood, Keith D.
    FCCM 2006: 14TH ANNUAL IEEE SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES, PROCEEDINGS, 2006, : 349 - +