Quantized Training of Gradient Boosting Decision Trees

被引:0
|
作者
Shi, Yu [1 ]
Ke, Guolin [2 ]
Chen, Zhuoming [3 ]
Zheng, Shuxin [1 ]
Liu, Tie-Yan [1 ]
机构
[1] Microsoft Res, Redmond, WA 98052 USA
[2] DP Technol, Beijing, Peoples R China
[3] Tsinghua Univ, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years have witnessed significant success in Gradient Boosting Decision Trees (GBDT) for a wide range of machine learning applications. Generally, a consensus about GBDT's training algorithms is gradients and statistics are computed based on high-precision floating points. In this paper, we investigate an essentially important question which has been largely ignored by the previous literature - how many bits are needed for representing gradients in training GBDT? To solve this mystery, we propose to quantize all the high-precision gradients in a very simple yet effective way in the GBDT's training algorithm. Surprisingly, both our theoretical analysis and empirical studies show that the necessary precisions of gradients without hurting any performance can be quite low, e.g., 2 or 3 bits. With low-precision gradients, most arithmetic operations in GBDT training can be replaced by integer operations of 8, 16, or 32 bits. Promisingly, these findings may pave the way for much more efficient training of GBDT from several aspects: (1) speeding up the computation of gradient statistics in histograms; (2) compressing the communication cost of high-precision statistical information during distributed training; (3) the inspiration of utilization and development of hardware architectures which well support low-precision computation for GBDT training. Benchmarked on CPUs, GPUs, and distributed clusters, we observe up to 2x speedup of our simple quantization strategy compared with SOTA GBDT systems on extensive datasets, demonstrating the effectiveness and potential of the low-precision training of GBDT. The code will be released to the official repository of LightGBM.(4)
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Booster: An Accelerator for Gradient Boosting Decision Trees Training and Inference
    He, Mingxuan
    Thottethodi, Mithuna
    Vijaykumar, T. N.
    2022 IEEE 36TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2022), 2022, : 1051 - 1062
  • [2] Machine Unlearning in Gradient Boosting Decision Trees
    Lin, Huawei
    Chung, Jun Woo
    Lao, Yingjie
    Zhao, Weijie
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 1374 - 1383
  • [3] Label Aggregation of Gradient Boosting Decision Trees
    Xiang, X. C.
    Zhang, H. X.
    Xia, S. T.
    PROCEEDINGS OF 2020 2ND INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND MACHINE VISION AND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION AND MACHINE LEARNING, IPMV 2020, 2020, : 140 - 145
  • [4] FPGA Accelerator for Gradient Boosting Decision Trees
    Alcolea, Adrian
    Resano, Javier
    ELECTRONICS, 2021, 10 (03) : 1 - 15
  • [5] On Incremental Learning for Gradient Boosting Decision Trees
    Zhang, Chongsheng
    Zhang, Yuan
    Shi, Xianjin
    Almpanidis, George
    Fan, Gaojuan
    Shen, Xiajiong
    NEURAL PROCESSING LETTERS, 2019, 50 (01) : 957 - 987
  • [6] Gradient Boosting Decision Trees for Echocardiogram Images
    de Melo, Vinicius Veloso
    Ushizima, Daniela Mayumi
    Baracho, Salety Ferreira
    Coelho, Regina Celia
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [7] Practical Federated Gradient Boosting Decision Trees
    Li, Qinbin
    Wen, Zeyi
    He, Bingsheng
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 4642 - 4649
  • [8] On Incremental Learning for Gradient Boosting Decision Trees
    Chongsheng Zhang
    Yuan Zhang
    Xianjin Shi
    George Almpanidis
    Gaojuan Fan
    Xiajiong Shen
    Neural Processing Letters, 2019, 50 : 957 - 987
  • [9] Speeding up Boosting decision trees training
    Zheng, Chao
    Wei, Zhenzhong
    AOPC 2015: IMAGE PROCESSING AND ANALYSIS, 2015, 9675
  • [10] Multi-Layered Gradient Boosting Decision Trees
    Feng, Ji
    Yu, Yang
    Zhou, Zhi-Hua
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31