On Embeddings for Numerical Features in Tabular Deep Learning

被引:0
|
作者
Gorishniy, Yury [1 ]
Rubachev, Ivan [2 ]
Babenko, Artem [1 ]
机构
[1] Yandex, Moscow, Russia
[2] HSE, Yandex, Moscow, Russia
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, Transformer-like deep architectures have shown strong performance on tabular data problems. Unlike traditional models, e.g., MLP, these architectures map scalar values of numerical features to high-dimensional embeddings before mixing them in the main backbone. In this work, we argue that embeddings for numerical features are an underexplored degree of freedom in tabular DL, which allows constructing more powerful DL models and competing with gradient boosted decision trees (GBDT) on some GBDT-friendly benchmarks (that is, where GBDT outperforms conventional DL models). We start by describing two conceptually different approaches to building embedding modules: the first one is based on a piecewise linear encoding of scalar values, and the second one utilizes periodic activations. Then, we empirically demonstrate that these two approaches can lead to significant performance boosts compared to the embeddings based on conventional blocks such as linear layers and ReLU activations. Importantly, we also show that embedding numerical features is beneficial for many backbones, not only for Transformers. Specifically, after proper embeddings, simple MLP-like models can perform on par with the attention-based architectures. Overall, we highlight embeddings for numerical features as an important design aspect with good potential for further improvements in tabular DL. The source code is available at https://github.com/Yura52/tabular-dl-num-embeddings.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] High dimensional, tabular deep learning with an auxiliary knowledge graph
    Ruiz, Camilo
    Ren, Hongyu
    Huang, Kexin
    Leskovec, Jure
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [22] Deep clustering of tabular data by weighted Gaussian distribution learning
    Rabbani, Shourav B.
    Medri, Ivan, V
    Samad, Manar D.
    NEUROCOMPUTING, 2025, 623
  • [23] Contextual word embeddings for tabular data search and integration
    Pilaluisa, Jose
    Tomas, David
    Navarro-Colorado, Borja
    Mazon, Jose-Norberto
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (13): : 9319 - 9333
  • [24] COMBINING DEEP EMBEDDINGS OF ACOUSTIC AND ARTICULATORY FEATURES FOR SPEAKER IDENTIFICATION
    Hong, Qian-Bei
    Wu, Chung-Hsien
    Wang, Hsin-Min
    Huang, Chien-Lin
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7589 - 7593
  • [25] How to Aggregate Acoustic Delta Features for Deep Speaker Embeddings
    Kim, Youngsam
    Roh, Jong-hyuk
    Cho, Kwantae
    Cho, Sangrae
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1225 - 1229
  • [26] Contextual word embeddings for tabular data search and integration
    José Pilaluisa
    David Tomás
    Borja Navarro-Colorado
    Jose-Norberto Mazón
    Neural Computing and Applications, 2023, 35 : 9319 - 9333
  • [27] Job and Employee Embeddings: A Joint Deep Learning Approach
    Liu, Hao
    Ge, Yong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (07) : 7056 - 7067
  • [28] Deep MinCut: Learning Node Embeddings by Detecting Communities
    Duong, Chi Thang
    Nguyen, Thanh Tam
    Hoang, Trung-Dung
    Yin, Hongzhi
    Weidlich, Matthias
    Nguyen, Quoc Viet Hung
    PATTERN RECOGNITION, 2022, 134
  • [29] LEARNING CONVOLUTIONAL NEURAL NETWORKS WITH DEEP PART EMBEDDINGS
    Gupta, Nitin
    Mujumdar, Shashank
    Agarwal, Prerna
    Jain, Abhinav
    Mehta, Sameep
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2037 - 2041
  • [30] Deep learning for universal linear embeddings of nonlinear dynamics
    Lusch, Bethany
    Kutz, J. Nathan
    Brunton, Steven L.
    NATURE COMMUNICATIONS, 2018, 9