Optimal Architecture of Floating-Point Arithmetic for Neural Network Training Processors

被引:11
|
作者
Junaid, Muhammad [1 ]
Arslan, Saad [2 ]
Lee, TaeGeon [1 ]
Kim, HyungWon [1 ]
机构
[1] Chungbuk Natl Univ, Coll Elect & Comp Engn, Dept Elect, Cheongju 28644, South Korea
[2] COMSATS Univ Islamabad, Dept Elect & Comp Engn, Pk Rd, Islamabad 45550, Pakistan
基金
新加坡国家研究基金会;
关键词
floating-points; IEEE; 754; convolutional neural network (CNN); MNIST dataset; ACCELERATOR;
D O I
10.3390/s22031230
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The convergence of artificial intelligence (AI) is one of the critical technologies in the recent fourth industrial revolution. The AIoT (Artificial Intelligence Internet of Things) is expected to be a solution that aids rapid and secure data processing. While the success of AIoT demanded low-power neural network processors, most of the recent research has been focused on accelerator designs only for inference. The growing interest in self-supervised and semi-supervised learning now calls for processors offloading the training process in addition to the inference process. Incorporating training with high accuracy goals requires the use of floating-point operators. The higher precision floating-point arithmetic architectures in neural networks tend to consume a large area and energy. Consequently, an energy-efficient/compact accelerator is required. The proposed architecture incorporates training in 32 bits, 24 bits, 16 bits, and mixed precisions to find the optimal floating-point format for low power and smaller-sized edge device. The proposed accelerator engines have been verified on FPGA for both inference and training of the MNIST image dataset. The combination of 24-bit custom FP format with 16-bit Brain FP has achieved an accuracy of more than 93%. ASIC implementation of this optimized mixed-precision accelerator using TSMC 65nm reveals an active area of 1.036 x 1.036 mm(2) and energy consumption of 4.445 mu J per training of one image. Compared with 32-bit architecture, the size and the energy are reduced by 4.7 and 3.91 times, respectively. Therefore, the CNN structure using floating-point numbers with an optimized data path will significantly contribute to developing the AIoT field that requires a small area, low energy, and high accuracy.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Training and Inference using Approximate Floating-Point Arithmetic for Energy Efficient Spiking Neural Network Processors
    Kwak, Myeongjin
    Lee, Jungwon
    Seo, Hyoju
    Sung, Mingyu
    Kim, Yongtae
    2021 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2021,
  • [2] Multiple precision floating-point arithmetic on SIMD processors
    van der Hoeven, Joris
    2017 IEEE 24TH SYMPOSIUM ON COMPUTER ARITHMETIC (ARITH), 2017, : 2 - 9
  • [3] OPTIMAL CHOICE OF BASIS FOR A FLOATING-POINT ARITHMETIC
    KREIFELTS, T
    COMPUTING, 1973, 11 (04) : 353 - 363
  • [4] OPTIMIZATION OF REAL NONRECURSIVE PROCESSORS IMPLEMENTED IN FLOATING-POINT ARITHMETIC
    RADONIA, PJ
    LECTURE NOTES IN CONTROL AND INFORMATION SCIENCES, 1990, 143 : 848 - 857
  • [5] Floating-point arithmetic
    Boldo, Sylvie
    Jeannerod, Claude-Pierre
    Melquiond, Guillaume
    Muller, Jean-Michel
    ACTA NUMERICA, 2023, 32 : 203 - 290
  • [6] Efficient Emulation of Floating-Point Arithmetic on Fixed-Point SIMD Processors
    Gerlach, Lukas
    Paya-Vaya, Guillermo
    Blume, Holger
    2016 IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2016, : 254 - 259
  • [7] Energy-Efficient Floating-Point Arithmetic for Digital Signal Processors
    Gilani, Syed Zohaib
    Kim, Nam Sung
    Schulte, Michael
    2011 CONFERENCE RECORD OF THE FORTY-FIFTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS (ASILOMAR), 2011, : 1823 - 1827
  • [8] ASIC Design of Nanoscale Artificial Neural Networks for Inference/Training by Floating-Point Arithmetic
    Niknia, Farzad
    Wang, Ziheng
    Liu, Shanshan
    Reviriego, Pedro
    Louri, Ahmed
    Lombardi, Fabrizio
    IEEE TRANSACTIONS ON NANOTECHNOLOGY, 2024, 23 : 208 - 216
  • [9] ROUNDINGS IN FLOATING-POINT ARITHMETIC
    YOHE, JM
    IEEE TRANSACTIONS ON COMPUTERS, 1973, C 22 (06) : 577 - 586
  • [10] Sabrewing: A Lightweight Architecture for Combined Floating-Point and Integer Arithmetic
    Bruintjes, Tom M.
    Walters, Karel H. G.
    Gerez, Sabih H.
    Molenkamp, Bert
    Smit, Gerard J. M.
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2012, 8 (04)