Gradient-Boosted Based Structured and Unstructured Learning

被引:0
|
作者
Gavito, Andrea Trevino [1 ]
Klabjan, Diego [1 ]
Utke, Jean [2 ]
机构
[1] Northwestern Univ, Evanston, IL 60208 USA
[2] Allstate Insurance Co, Northbrook, IL USA
关键词
Deep learning; Multimodal learning; Gradient boosting;
D O I
10.1007/978-3-031-44213-1_37
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose two frameworks to deal with problem settings in which both structured and unstructured data are available. Structured data problems are best solved by traditional machine learning models such as boosting and tree-based algorithms, whereas deep learning has been widely applied to problems dealing with images, text, audio, and other unstructured data sources. However, for the setting in which both structured and unstructured data are accessible, it is not obvious what the best modeling approach is to enhance performance on both data sources simultaneously. Our proposed frameworks allow joint learning on both kinds of data by integrating the paradigms of boosting models and deep neural networks. The first framework, the boosted-feature-vector deep learning network, learns features from the structured data using gradient boosting and combines them with embeddings from unstructured data via a two-branch deep neural network. Secondly, the two-weak-learner boosting framework extends the boosting paradigm to the setting with two input data sources. We present and compare first- and second-order methods of this framework. Our experimental results on both public and real-world datasets show performance gains achieved by the frameworks over selected baselines by magnitudes of 0.1%-4.7%.
引用
收藏
页码:439 / 451
页数:13
相关论文
共 50 条
  • [1] Gradient-boosted equivalent sources
    Soler, Santiago R.
    Uieda, Leonardo
    GEOPHYSICAL JOURNAL INTERNATIONAL, 2021, 227 (03) : 1768 - 1783
  • [2] Instance-Based Uncertainty Estimation for Gradient-Boosted Regression Trees
    Brophy, Jonathan
    Lowd, Daniel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [3] Adversarial Training of Gradient-Boosted Decision Trees
    Calzavara, Stefano
    Lucchese, Claudio
    Tolomei, Gabriele
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2429 - 2432
  • [4] Classification and Recognition of Building Appearance Based on Optimized Gradient-Boosted Decision Tree Algorithm
    Hu, Mengting
    Guo, Lingxiang
    Liu, Jing
    Song, Yuxuan
    SENSORS, 2023, 23 (11)
  • [5] Gradient-Boosted Generalized Linear Models for Conditional Vine Copulas
    Jobst, David
    Moeller, Annette
    Gross, Juergen
    ENVIRONMETRICS, 2024,
  • [6] Dynamic Streamflow Simulation via Online Gradient-Boosted Regression Tree
    Zhang, Heng
    Yang, Qinli
    Shao, Junming
    Wang, Guoqing
    JOURNAL OF HYDROLOGIC ENGINEERING, 2019, 24 (10)
  • [7] GBDT-MO: Gradient-Boosted Decision Trees for Multiple Outputs
    Zhang, Zhendong
    Jung, Cheolkon
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (07) : 3156 - 3167
  • [8] Forecasting PM2.5 Concentration Using Gradient-Boosted Regression Tree with CNN Learning Model
    Usha Ruby, A.
    Chandran, J. George Chellin
    Theerthagiri, Prasannavenkatesan
    Patil, Renuka
    Chaithanya, B. N.
    Jain, T. J. Swasthika
    OPTICAL MEMORY AND NEURAL NETWORKS, 2024, 33 (01) : 86 - 96
  • [9] Evaluating soccer match prediction models: a deep learning approach and feature optimization for gradient-boosted trees
    Yeung, Calvin
    Bunker, Rory
    Umemoto, Rikuhei
    Fujii, Keisuke
    MACHINE LEARNING, 2024, 113 (10) : 7541 - 7564
  • [10] Forecasting PM2.5 Concentration Using Gradient-Boosted Regression Tree with CNN Learning Model
    A. Usha Ruby
    J. George Chellin Chandran
    Prasannavenkatesan Theerthagiri
    Renuka Patil
    B. N. Chaithanya
    T. J. Swasthika Jain
    Optical Memory and Neural Networks, 2024, 33 : 86 - 96