Quantized Training of Gradient Boosting Decision Trees

被引:0
|
作者
Shi, Yu [1 ]
Ke, Guolin [2 ]
Chen, Zhuoming [3 ]
Zheng, Shuxin [1 ]
Liu, Tie-Yan [1 ]
机构
[1] Microsoft Res, Redmond, WA 98052 USA
[2] DP Technol, Beijing, Peoples R China
[3] Tsinghua Univ, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years have witnessed significant success in Gradient Boosting Decision Trees (GBDT) for a wide range of machine learning applications. Generally, a consensus about GBDT's training algorithms is gradients and statistics are computed based on high-precision floating points. In this paper, we investigate an essentially important question which has been largely ignored by the previous literature - how many bits are needed for representing gradients in training GBDT? To solve this mystery, we propose to quantize all the high-precision gradients in a very simple yet effective way in the GBDT's training algorithm. Surprisingly, both our theoretical analysis and empirical studies show that the necessary precisions of gradients without hurting any performance can be quite low, e.g., 2 or 3 bits. With low-precision gradients, most arithmetic operations in GBDT training can be replaced by integer operations of 8, 16, or 32 bits. Promisingly, these findings may pave the way for much more efficient training of GBDT from several aspects: (1) speeding up the computation of gradient statistics in histograms; (2) compressing the communication cost of high-precision statistical information during distributed training; (3) the inspiration of utilization and development of hardware architectures which well support low-precision computation for GBDT training. Benchmarked on CPUs, GPUs, and distributed clusters, we observe up to 2x speedup of our simple quantization strategy compared with SOTA GBDT systems on extensive datasets, demonstrating the effectiveness and potential of the low-precision training of GBDT. The code will be released to the official repository of LightGBM.(4)
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Investigating the contributors to hit-and-run crashes using gradient boosting decision trees
    Han, Baorui
    Huang, Haibo
    Li, Gen
    Jiang, Chenming
    Yang, Zhen
    Zhu, Zhenjun
    PLOS ONE, 2025, 20 (01):
  • [42] Gradient boosting decision trees for cyber security threats detection based on network events logs
    Quang Hieu Vu
    Ruta, Dymitr
    Cen, Ling
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 5921 - 5928
  • [43] Histogram-Based Algorithm for Building Gradient Boosting Ensembles of Piecewise Linear Decision Trees
    Guryanov, Aleksei
    ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS, AIST 2019, 2019, 11832 : 39 - 50
  • [44] Analysis of the State of High-Voltage Current Transformers Based on Gradient Boosting on Decision Trees
    Khalyasmaa, Alexandra I.
    Senyuk, Mihail D.
    Eroshenko, Stanislav A.
    IEEE TRANSACTIONS ON POWER DELIVERY, 2021, 36 (04) : 2154 - 2163
  • [45] Predictive value of gradient boosting decision trees for postoperative atelectasis complications in patients with pulmonary destruction
    Tang, Zhongming
    Tang, Jifu
    Liu, Wei
    Chen, Guoqiang
    Feng, Chenggang
    Zhang, Aiping
    AMERICAN JOURNAL OF TRANSLATIONAL RESEARCH, 2024, 16 (07): : 2864 - 2876
  • [46] User Clustering for MIMO NOMA via Classifier Chains and Gradient-Boosting Decision Trees
    Ben Issaid, Chaouki
    Anton-Haro, Carles
    Mestre, Xavier
    Alouini, Mohamed-Slim
    IEEE ACCESS, 2020, 8 (08): : 211411 - 211421
  • [47] Forecasting Nickel futures price based on the empirical wavelet transform and gradient boosting decision trees
    Gu, Qinghua
    Chang, Yinxin
    Xiong, Naixue
    Chen, Lu
    APPLIED SOFT COMPUTING, 2021, 109
  • [48] Constraint learning based gradient boosting trees
    Israeli, Abraham
    Rokach, Lior
    Shabtai, Asaf
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 128 : 287 - 300
  • [49] Predicting Phenotypes From High-Dimensional Genomes Using Gradient Boosting Decision Trees
    Yu, Tingxi
    Wang, Li
    Zhang, Wuping
    Xing, Guofang
    Han, Jiwan
    Li, Fuzhong
    Cao, Chunqing
    IEEE ACCESS, 2022, 10 : 48126 - 48140
  • [50] Asynch-SGBDT: Train Stochastic Gradient Boosting Decision Trees in an Asynchronous Parallel Manner
    Cheng, Daning
    Li, Shigang
    Zhang, Yunquan
    2023 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM, IPDPS, 2023, : 256 - 267