Optimally Approximated and Unbiased Floating-Point Multiplier with Runtime Configurability

被引:0
|
作者
Chen, Chuangtao [2 ]
Yang, Sen [1 ]
Qian, Weikang [4 ]
Imani, Mohsen [5 ]
Yin, Xunzhao [1 ]
Zhuo, Cheng [1 ,3 ]
机构
[1] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou, Peoples R China
[2] Zhejiang Univ, Coll Elect Engn, Hangzhou, Peoples R China
[3] Fudan Univ, Sch Microelect, ASIC & Syst Key Lab, Shanghai, Peoples R China
[4] Shanghai Jiao Tong Univ, Univ Michigan Shanghai Jiao Tong Univ Joint Inst, Shanghai, Peoples R China
[5] Univ Calif Irvine, Dept Comp Sci & Engn, Irvine, CA USA
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Approximate computing is a promising alternative to improve energy efficiency for IoT devices on the edge. This work proposes an optimally approximated and unbiased floating-point approximate multiplier with runtime configurability. We provide a theoretically sound formulation that turns multiplication approximation to an optimization problem. With the formulation and findings, a multilevel architecture is proposed to easily incorporate runtime configurability and module execution parallelism. Finally, an optimization scheme is applied to improve the area, making it linearly dependent on the precision, instead of quadratically or exponentially as in prior work. In addition to the optimal approximation and configurability, the proposed design has an efficient circuit implementation that uses inversion, shift and addition instead of complex arithmetic operations. When compared to the prior state-of-the-art approximate floating-point multiplier, ApproxLP [30], the proposed design outperforms in all aspects including accuracy, area, and delay. By replacing the regular full-precision multiplier in GPU, the proposed design can improve the energy efficiency for various edge computing tasks. Even with Level 1 approximation, the proposed design improves energy efficiency up to 122x for machine learning on C1FAR-10, with almost negligible accuracy loss.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Small Logarithmic Floating-Point Multiplier Based on FPGA and Its Application on MobileNet
    Xiong, Botao
    Fan, Sheng
    He, Xintong
    Xu, Tu
    Chang, Yuchun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (12) : 5119 - 5123
  • [42] A Binary Integer Decimal-based multiplier for Decimal Floating-Point arithmetic
    Gonzalez-Navarro, Sonia
    Tsen, Charles
    Schulte, NEchael
    CONFERENCE RECORD OF THE FORTY-FIRST ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, VOLS 1-5, 2007, : 353 - +
  • [43] Fault-Tolerant Floating-Point Multiplier Design for Mission Critical Systems
    Kumar, Sakali Raghavendra
    Veeramachaneni, Sreehari
    Mahammad, Noor Sk
    PROCEEDINGS OF THE 37TH INTERNATIONAL CONFERENCE ON VLSI DESIGN, VLSID 2024 AND 23RD INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS, ES 2024, 2024, : 678 - 683
  • [44] A Multi-Format Floating-Point Multiplier for Power-Efficient Operations
    Nannarelli, Alberto
    2017 30TH IEEE INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE (SOCC), 2017, : 351 - 356
  • [45] An efficient multi-format low-precision floating-point multiplier
    Kermani, Hadis Ahmadpour
    Zarandi, Azadeh Alsadat Emrani
    SUSTAINABLE COMPUTING-INFORMATICS & SYSTEMS, 2024, 41
  • [46] Floating-Point Division Algorithms for an x86 Microprocessor with a Rectangular Multiplier
    Schulte, Michael J.
    Tan, Dimitri
    Lemonds, Carl E.
    2007 IEEE INTERNATIONAL CONFERENCE ON COMPUTER DESIGN, VOLS, 1 AND 2, 2007, : 304 - +
  • [47] Floating-point arithmetic
    Boldo, Sylvie
    Jeannerod, Claude-Pierre
    Melquiond, Guillaume
    Muller, Jean-Michel
    ACTA NUMERICA, 2023, 32 : 203 - 290
  • [48] On floating-point summation
    Espelid, TO
    SIAM REVIEW, 1995, 37 (04) : 603 - 607
  • [49] FLOATING-POINT COMPUTATION
    STERBENZ, P
    TRANSACTIONS OF THE NEW YORK ACADEMY OF SCIENCES, 1974, 36 (06): : 591 - 591
  • [50] Floating-point tricks
    Blinn, JF
    IEEE COMPUTER GRAPHICS AND APPLICATIONS, 1997, 17 (04) : 80 - 84