SigOpt Mulch: An intelligent system for AutoML of gradient boosted trees

被引:1
|
作者
Sorokin, Aleksei [1 ]
Zhu, Xinran [2 ]
Lee, Eric Hans [3 ]
Cheng, Bolong [3 ]
机构
[1] IIT, Chicago, IL USA
[2] Cornell Univ, Ithaca, NY USA
[3] SigOpt Intel Co, San Francisco, CA 94104 USA
关键词
Automated machine learning (autoML); Hyperparameter optimization (HPO); Gradient boosted trees; EFFICIENT;
D O I
10.1016/j.knosys.2023.110604
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gradient boosted trees (GBTs) are ubiquitous models used by researchers, machine learning (ML) practitioners, and data scientists because of their robust performance, interpretable behavior, and ease-of-use. One critical challenge in training GBTs is the tuning of their hyperparameters. In practice, selecting these hyperparameters is often done manually. Recently, the ML community has advocated for tuning hyperparameters through black-box optimization and developed state-of-the-art systems to do so. However, applying such systems to tune GBTs suffers from two drawbacks. First, these systems are not model-aware, rather they are designed to apply to a generic model; this leaves significant optimization performance on the table. Second, using these systems requires domain knowledge such as the choice of hyperparameter search space, which is an antithesis to the automatic experimentation that black-box optimization aims to provide. In this paper, we present SigOpt Mulch, a model-aware hyperparameter tuning system specifically designed for automated tuning of GBTs that provides two improvements over existing systems. First, Mulch leverages powerful techniques in metalearning and multifidelity optimization to perform model-aware hyperparameter optimization. Second, it automates the process of learning performant hyperparameters by making intelligent decisions about the optimization search space, thus reducing the need for user domain knowledge. These innovations allow Mulch to identify good GBT hyperparameters far more efficiently-and in a more seamless and user-friendly way-than existing black-box hyperparameter tuning systems. (c) 2023 Published by Elsevier B.V.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Gradient Boosted Trees for Corrective Learning
    Oguz, Baris U.
    Shinohara, Russell T.
    Yushkevich, Paul A.
    Oguz, Ipek
    MACHINE LEARNING IN MEDICAL IMAGING (MLMI 2017), 2017, 10541 : 203 - 211
  • [2] Counting People using Gradient Boosted Trees
    Zhou, Bingyin
    Lu, Ming
    Wang, Yonggang
    2016 IEEE INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC), 2016, : 391 - 395
  • [3] Robust Supply Chains with Gradient Boosted Trees
    Mahato, Pradeep K.
    Narayan, Apurva
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 2633 - 2639
  • [4] Gradient boosted trees for evolving data streams
    Nuwan Gunasekara
    Bernhard Pfahringer
    Heitor Gomes
    Albert Bifet
    Machine Learning, 2024, 113 : 3325 - 3352
  • [5] GRADIENT BOOSTED DECISION TREES FOR LITHOLOGY CLASSIFICATION
    Dev, Vikrant A.
    Eden, Mario R.
    PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON FOUNDATIONS OF COMPUTER-AIDED PROCESS DESIGN, 2019, 47 : 113 - 118
  • [6] Block-distributed Gradient Boosted Trees
    Vasiloudis, Theodore
    Cho, Hyunsu
    Bostrom, Henrik
    PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 1025 - 1028
  • [7] Gradient boosted trees for evolving data streams
    Gunasekara, Nuwan
    Pfahringer, Bernhard
    Gomes, Heitor
    Bifet, Albert
    MACHINE LEARNING, 2024, 113 (05) : 3325 - 3352
  • [8] Leaves on trees: identifying halo stars with extreme gradient boosted trees
    Veljanoski, Jovan
    Helmi, Amina
    Breddels, Maarten
    Posti, Lorenzo
    ASTRONOMY & ASTROPHYSICS, 2018, 621
  • [9] Syntax Description Synthesis Using Gradient Boosted Trees
    Astashkin, Arseny
    Chuvilin, Kirill
    PROCEEDINGS OF THE 20TH CONFERENCE OF OPEN INNOVATIONS ASSOCIATION (FRUCT 2017), 2017, : 32 - 39
  • [10] Adversarial Training of Gradient-Boosted Decision Trees
    Calzavara, Stefano
    Lucchese, Claudio
    Tolomei, Gabriele
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2429 - 2432