SigOpt Mulch: An intelligent system for AutoML of gradient boosted trees

被引:1
|
作者
Sorokin, Aleksei [1 ]
Zhu, Xinran [2 ]
Lee, Eric Hans [3 ]
Cheng, Bolong [3 ]
机构
[1] IIT, Chicago, IL USA
[2] Cornell Univ, Ithaca, NY USA
[3] SigOpt Intel Co, San Francisco, CA 94104 USA
关键词
Automated machine learning (autoML); Hyperparameter optimization (HPO); Gradient boosted trees; EFFICIENT;
D O I
10.1016/j.knosys.2023.110604
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Gradient boosted trees (GBTs) are ubiquitous models used by researchers, machine learning (ML) practitioners, and data scientists because of their robust performance, interpretable behavior, and ease-of-use. One critical challenge in training GBTs is the tuning of their hyperparameters. In practice, selecting these hyperparameters is often done manually. Recently, the ML community has advocated for tuning hyperparameters through black-box optimization and developed state-of-the-art systems to do so. However, applying such systems to tune GBTs suffers from two drawbacks. First, these systems are not model-aware, rather they are designed to apply to a generic model; this leaves significant optimization performance on the table. Second, using these systems requires domain knowledge such as the choice of hyperparameter search space, which is an antithesis to the automatic experimentation that black-box optimization aims to provide. In this paper, we present SigOpt Mulch, a model-aware hyperparameter tuning system specifically designed for automated tuning of GBTs that provides two improvements over existing systems. First, Mulch leverages powerful techniques in metalearning and multifidelity optimization to perform model-aware hyperparameter optimization. Second, it automates the process of learning performant hyperparameters by making intelligent decisions about the optimization search space, thus reducing the need for user domain knowledge. These innovations allow Mulch to identify good GBT hyperparameters far more efficiently-and in a more seamless and user-friendly way-than existing black-box hyperparameter tuning systems. (c) 2023 Published by Elsevier B.V.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Boulevard: Regularized Stochastic Gradient Boosted Trees and Their Limiting Distribution
    Zhou, Yichen
    Hooker, Giles
    Journal of Machine Learning Research, 2022, 23
  • [22] Estimation of inorganic crystal densities using gradient boosted trees
    Zhao, Jesse
    FRONTIERS IN MATERIALS, 2022, 9
  • [23] Estimation of the masses in the local group by gradient boosted decision trees
    Carlesi, Edoardo
    Hoffman, Yehuda
    Libeskind, Noam, I
    MONTHLY NOTICES OF THE ROYAL ASTRONOMICAL SOCIETY, 2022, 513 (02) : 2385 - 2393
  • [24] Waist circumference prediction for epidemiological research using gradient boosted trees
    Zhou, Weihong
    Eckler, Spencer
    Barszczyk, Andrew
    Waese-Perlman, Alex
    Wang, Yingjie
    Gu, Xiaoping
    Feng, Zhong-Ping
    Peng, Yuzhu
    Lee, Kang
    BMC MEDICAL RESEARCH METHODOLOGY, 2021, 21 (01)
  • [25] Gradient Boosted Trees and Denoising Autoencoder to Correct Numerical Wave Forecasts
    Yanchin, Ivan
    Soares, C. Guedes
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2024, 12 (09)
  • [26] Modeling Tick Populations: An Ecological Test Case for Gradient Boosted Trees
    Manley, William
    Tran, Tam
    Prusinski, Melissa
    Brisson, Dustin
    PEER COMMUNITY JOURNAL, 2023, 3
  • [27] Formation lithology classification using scalable gradient boosted decision trees
    Dev, Vikrant A.
    Eden, Mario R.
    COMPUTERS & CHEMICAL ENGINEERING, 2019, 128 : 392 - 404
  • [28] Wind Ramp Event Prediction with Parallelized Gradient Boosted Regression Trees
    Gupta, Saurav
    Shrivastava, Nitin Anand
    Khosravi, Abbas
    Panigrahi, Bijaya Ketan
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 5296 - 5301
  • [29] Gradient boosted decision trees reveal nuances of auditory discrimination behavior
    Griffiths, Carla S.
    Lebert, Jules M.
    Sollini, Joseph
    Bizley, Jennifer K.
    PLOS COMPUTATIONAL BIOLOGY, 2024, 20 (04)
  • [30] GB-CENT: Gradient Boosted Categorical Embedding and Numerical Trees
    Zhao, Qian
    Shi, Yue
    Hong, Liangjie
    PROCEEDINGS OF THE 26TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'17), 2017, : 1311 - 1319