On Embeddings for Numerical Features in Tabular Deep Learning

被引:0
|
作者
Gorishniy, Yury [1 ]
Rubachev, Ivan [2 ]
Babenko, Artem [1 ]
机构
[1] Yandex, Moscow, Russia
[2] HSE, Yandex, Moscow, Russia
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, Transformer-like deep architectures have shown strong performance on tabular data problems. Unlike traditional models, e.g., MLP, these architectures map scalar values of numerical features to high-dimensional embeddings before mixing them in the main backbone. In this work, we argue that embeddings for numerical features are an underexplored degree of freedom in tabular DL, which allows constructing more powerful DL models and competing with gradient boosted decision trees (GBDT) on some GBDT-friendly benchmarks (that is, where GBDT outperforms conventional DL models). We start by describing two conceptually different approaches to building embedding modules: the first one is based on a piecewise linear encoding of scalar values, and the second one utilizes periodic activations. Then, we empirically demonstrate that these two approaches can lead to significant performance boosts compared to the embeddings based on conventional blocks such as linear layers and ReLU activations. Importantly, we also show that embedding numerical features is beneficial for many backbones, not only for Transformers. Specifically, after proper embeddings, simple MLP-like models can perform on par with the attention-based architectures. Overall, we highlight embeddings for numerical features as an important design aspect with good potential for further improvements in tabular DL. The source code is available at https://github.com/Yura52/tabular-dl-num-embeddings.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Learning Tabular Embeddings at Web Scale
    Pavia, Sophie
    Shams, Montasir
    Khan, Rituparna
    Pyayt, Anna
    Gubanov, Michael
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 4936 - 4944
  • [2] Tabular and Deep Learning for the Whittle Index
    Relano, Francisco Robledo
    Borkar, Vivek
    Ayesta, Urtzi
    Avrachenko, Konstan
    ACM TRANSACTIONS ON MODELING AND PERFORMANCE EVALUATION OF COMPUTING SYSTEMS, 2024, 9 (03)
  • [3] Leveraging Periodicity for Tabular Deep Learning
    Rizzo, Matteo
    Ayyurek, Ebru
    Albarelli, Andrea
    Gasparetto, Andrea
    ELECTRONICS, 2025, 14 (06):
  • [4] Regularization Learning Networks: Deep Learning for Tabular Datasets
    Shavitt, Ira
    Segal, Eran
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [5] Revisiting Deep Learning Models for Tabular Data
    Gorishniy, Yury
    Rubachev, Ivan
    Khrulkov, Valentin
    Babenko, Artem
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] LocalGLMnet: interpretable deep learning for tabular data
    Richman, Ronald
    Wuethrich, Mario, V
    SCANDINAVIAN ACTUARIAL JOURNAL, 2023, 2023 (01) : 71 - 95
  • [7] Recent deep learning methods for tabular data
    Hwang, Yejin
    Song, Jongwoo
    COMMUNICATIONS FOR STATISTICAL APPLICATIONS AND METHODS, 2023, 30 (02) : 215 - 226
  • [8] Is Deep Learning on Tabular Data Enough? An Assessment
    Fayaz, Sheikh Amir
    Zaman, Majid
    Kaul, Sameer
    Butt, Muheet Ahmed
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (04) : 466 - 473
  • [9] Learning Deep Embeddings with Histogram Loss
    Ustinova, Evgeniya
    Lempitsky, Victor
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [10] Tabular data: Deep learning is not all you need
    Shwartz-Ziv, Ravid
    Armon, Amitai
    INFORMATION FUSION, 2022, 81 : 84 - 90