Quantized Training of Gradient Boosting Decision Trees

被引:0
|
作者
Shi, Yu [1 ]
Ke, Guolin [2 ]
Chen, Zhuoming [3 ]
Zheng, Shuxin [1 ]
Liu, Tie-Yan [1 ]
机构
[1] Microsoft Res, Redmond, WA 98052 USA
[2] DP Technol, Beijing, Peoples R China
[3] Tsinghua Univ, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent years have witnessed significant success in Gradient Boosting Decision Trees (GBDT) for a wide range of machine learning applications. Generally, a consensus about GBDT's training algorithms is gradients and statistics are computed based on high-precision floating points. In this paper, we investigate an essentially important question which has been largely ignored by the previous literature - how many bits are needed for representing gradients in training GBDT? To solve this mystery, we propose to quantize all the high-precision gradients in a very simple yet effective way in the GBDT's training algorithm. Surprisingly, both our theoretical analysis and empirical studies show that the necessary precisions of gradients without hurting any performance can be quite low, e.g., 2 or 3 bits. With low-precision gradients, most arithmetic operations in GBDT training can be replaced by integer operations of 8, 16, or 32 bits. Promisingly, these findings may pave the way for much more efficient training of GBDT from several aspects: (1) speeding up the computation of gradient statistics in histograms; (2) compressing the communication cost of high-precision statistical information during distributed training; (3) the inspiration of utilization and development of hardware architectures which well support low-precision computation for GBDT training. Benchmarked on CPUs, GPUs, and distributed clusters, we observe up to 2x speedup of our simple quantization strategy compared with SOTA GBDT systems on extensive datasets, demonstrating the effectiveness and potential of the low-precision training of GBDT. The code will be released to the official repository of LightGBM.(4)
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Gradient boosting decision trees to study laboratory and field performance in pavement management
    Berangi, Mohammadjavad
    Lontra, Bernardo Mota
    Anupam, Kumar
    Erkens, Sandra
    Van Vliet, Dave
    Snippe, Almar
    Moenielal, Mahesh
    COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2025, 40 (01) : 3 - 32
  • [22] Ensembling Learning Based Melanoma Classification Using Gradient Boosting Decision Trees
    Han, Yipeng
    Zheng, Xiaolu
    AIPR 2020: 2020 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND PATTERN RECOGNITION, 2020, : 104 - 109
  • [23] A mobile recommendation system based on Logistic Regression and Gradient Boosting Decision Trees
    Wang, Yaozheng
    Feng, Dawei
    Ii, Dongsheng
    Chen, Xinyuan
    Zhac, Yunxiang
    Niu, Xin
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 1896 - 1902
  • [24] Explainable Steel Quality Prediction System Based on Gradient Boosting Decision Trees
    Takalo-Mattila, Janne
    Heiskanen, Mikko
    Kyllonen, Vesa
    Maatta, Leena
    Bogdanoff, Agne
    IEEE ACCESS, 2022, 10 : 68099 - 68110
  • [25] Retrieval-Based Gradient Boosting Decision Trees for Disease Risk Assessment
    Ma, Handong
    Cao, Jiahang
    Fang, Yuchen
    Zhang, Weinan
    Sheng, Wenbo
    Zhang, Shaodian
    Yu, Yong
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 3468 - 3476
  • [26] Fast Gradient Boosting Decision Trees with Bit-Level Data Structures
    Devos, Laurens
    Meert, Wannes
    Davis, Jesse
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT I, 2020, 11906 : 590 - 606
  • [27] HYBRID MODEL FOR NETWORK ANOMALY DETECTION WITH GRADIENT BOOSTING DECISION TREES AND TABTRANSFORMER
    Xu, Xinyue
    Zheng, Xiaolu
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 8538 - 8542
  • [28] Automated formatting verification technique of paperwork based on the gradient boosting on decision trees
    Nasyrov, Nail
    Komarov, Mikhail
    Tartynskikh, Petr
    Gorlushkina, Nataliya
    9TH INTERNATIONAL YOUNG SCIENTISTS CONFERENCE IN COMPUTATIONAL SCIENCE, YSC2020, 2020, 178 : 365 - 374
  • [29] eFL-Boost: Efficient Federated Learning for Gradient Boosting Decision Trees
    Yamamoto, Fuki
    Ozawa, Seiichi
    Wang, Lihua
    IEEE ACCESS, 2022, 10 : 43954 - 43963
  • [30] Towards Fair and Decentralized Federated Learning System for Gradient Boosting Decision Trees
    Gao, Shiqi
    Li, Xianxian
    Shi, Zhenkui
    Liu, Peng
    Li, Chunpei
    SECURITY AND COMMUNICATION NETWORKS, 2022, 2022