Is Integer Arithmetic Enough for Deep Learning Training?

被引:0
|
作者
Ghaffari, Alireza [1 ]
Tahaei, Marzieh S. [1 ]
Tayaranian, Mohammadreza [1 ]
Asgharian, Masoud [2 ]
Nia, Vahid Partovi [1 ]
机构
[1] Montreal Res Ctr, Huawei Noahs Ark Lab, Montreal, PQ, Canada
[2] McGill Univ, Dept Math & Stat, Montreal, PQ, Canada
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The ever-increasing computational complexity of deep learning models makes their training and deployment difficult on various cloud and edge platforms. Replacing floating-point arithmetic with low-bit integer arithmetic is a promising approach to save energy, memory footprint, and latency of deep learning models. As such, quantization has attracted the attention of researchers in recent years. However, using integer numbers to form a fully functional integer training pipeline including forward pass, back-propagation, and stochastic gradient descent is not studied in detail. Our empirical and mathematical results reveal that integer arithmetic seems to be enough to train deep learning models. Unlike recent proposals, instead of quantization, we directly switch the number representation of computations. Our novel training method forms a fully integer training pipeline that does not change the trajectory of the loss and accuracy compared to floating-point, nor does it need any special hyper-parameter tuning, distribution adjustment, or gradient clipping. Our experimental results show that our proposed method is effective in a wide variety of tasks such as classification (including vision transformers), object detection, and semantic segmentation.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] NITI: Training Integer Neural Networks Using Integer-Only Arithmetic
    Wang, Maolin
    Rasoulinezhad, Seyedramin
    Leong, Philip H. W.
    So, Hayden K. -H.
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (11) : 3249 - 3261
  • [2] A Multi-level Parallel Integer/Floating-Point Arithmetic Architecture for Deep Learning Instructions
    Tan, Hongbing
    Zhang, Jing
    Huang, Libo
    He, Xiaowei
    Dong, Dezun
    Wang, Yongwen
    Xiao, Liquan
    EURO-PAR 2023: PARALLEL PROCESSING, 2023, 14100 : 260 - 274
  • [3] Learning Arithmetic Operations With A Multistep Deep Learning
    Nollet, Bastien
    Lefort, Mathieu
    Armetta, Frederic
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [4] Integer Is Enough: When Vertical Federated Learning Meets Rounding
    Qiu, Pengyu
    Pu, Yuwen
    Liu, Yongchao
    Liu, Wenyan
    Yue, Yun
    Zhu, Xiaowei
    Li, Lichun
    Li, Jinbao
    Ji, Shouling
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14704 - 14712
  • [5] Widening integer arithmetic
    Redwine, K
    Ramsey, N
    COMPILER CONSTRUCTION, PROCEEDINGS, 2004, 2985 : 232 - 249
  • [6] Modular Integer Arithmetic
    Schwarzweller, Christoph
    FORMALIZED MATHEMATICS, 2008, 16 (03): : 247 - 252
  • [7] Zeckendorf integer arithmetic
    Fenwick, P
    FIBONACCI QUARTERLY, 2003, 41 (05): : 405 - 413
  • [8] Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference
    Jacob, Benoit
    Kligys, Skirmantas
    Chen, Bo
    Zhu, Menglong
    Tang, Matthew
    Howard, Andrew
    Adam, Hartwig
    Kalenichenko, Dmitry
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 2704 - 2713
  • [9] Arithmetic learning in children: An fMRI training study
    Declercq, Merel
    Bellon, Elien
    Sahan, Muhammet Ikbal
    Fias, Wim
    De Smedt, Bert
    NEUROPSYCHOLOGIA, 2022, 169
  • [10] Is Distance Matrix Enough for Geometric Deep Learning?
    Li, Zian
    Wang, Xiyuan
    Huang, Yinan
    Zhang, Muhan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,