Robust Representation Learning via Sparse Attention Mechanism for Similarity Models

被引:0
|
作者
Ermilova, Alina [1 ]
Baramiia, Nikita [1 ]
Kornilov, Valerii [1 ]
Petrakov, Sergey [1 ]
Zaytsev, Alexey [1 ,2 ]
机构
[1] Skolkovo Inst Sci & Technol, Moscow 121205, Russia
[2] Sber, Risk Management, Moscow 121165, Russia
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Transformers; Oil insulation; Task analysis; Time series analysis; Meteorology; Training; Deep learning; Representation learning; efficient transformer; robust transformer; representation learning; similarity learning; TRANSFORMER;
D O I
10.1109/ACCESS.2024.3418779
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The attention-based models are widely used for time series data. However, due to the quadratic complexity of attention regarding input sequence length, the application of Transformers is limited by high resource demands. Moreover, their modifications for industrial time series need to be robust to missing or noisy values, which complicates the expansion of their application horizon. To cope with these issues, we introduce the class of efficient Transformers named Regularized Transformers (Reguformers). We implement the regularization technique inspired by the dropout ideas to improve robustness and reduce computational expenses without significantly modifying the pipeline. The focus in our experiments is on oil&gas data. For well-interval similarity task, our best Reguformer configuration reaches ROC AUC 0.97, which is comparable to Informer (0.978) and outperforms baselines: the previous LSTM model (0.934), the classical Transformer model (0.967), and three recent most promising modifications of the original Transformer, namely, Performer (0.949), LRformer (0.955), and DropDim (0.777). We also conduct the corresponding experiments on three additional datasets from different domains and obtain superior results. The increase in the quality of the best Reguformer relative to Transformer for different datasets varies from 3.7% to 9.6%, while the increase range relative to Informer is wider: from 1.7% to 18.4%.
引用
收藏
页码:97833 / 97850
页数:18
相关论文
共 50 条
  • [1] Robust Representation Learning via Perceptual Similarity Metrics
    Taghanaki, Saeid Asgari
    Choi, Kristy
    Khasahmadi, Amir
    Goyal, Anirudh
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7045 - 7056
  • [2] A Graph Representation Learning Algorithm Based on Attention Mechanism and Node Similarity
    Guo, Kun
    Wang, Deqin
    Huang, Jiangsheng
    Chen, Yuzhong
    Zhu, Zhihao
    Zheng, Jianning
    COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING, CHINESECSCW 2019, 2019, 1042 : 591 - 604
  • [3] LEARNING LOW RANK AND SPARSE MODELS VIA ROBUST AUTOENCODERS
    Pu, Jie
    Panagakis, Yannis
    Pantic, Maja
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 3192 - 3196
  • [4] Towards Robust Knowledge Tracing Models via k-Sparse Attention
    Huang, Shuyan
    Liu, Zitao
    Zhao, Xiangyu
    Luo, Weiqi
    Weng, Jian
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2441 - 2445
  • [5] Robust sparse coding via self-paced learning for data representation
    Feng, Xiaodong
    Wu, Sen
    INFORMATION SCIENCES, 2021, 546 : 448 - 468
  • [6] Robust Visual Tracking via Incremental Subspace Learning and Local Sparse Representation
    Yang, Guoliang
    Hu, Zhengwei
    Tang, Jun
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2018, 43 (02) : 627 - 636
  • [7] Robust Ship Tracking via Multi-view Learning and Sparse Representation
    Chen, Xinqiang
    Wang, Shengzheng
    Shi, Chaojian
    Wu, Huafeng
    Zhao, Jiansen
    Fu, Junjie
    JOURNAL OF NAVIGATION, 2019, 72 (01): : 176 - 192
  • [8] Robust Visual Tracking via Incremental Subspace Learning and Local Sparse Representation
    Guoliang Yang
    Zhengwei Hu
    Jun Tang
    Arabian Journal for Science and Engineering, 2018, 43 : 627 - 636
  • [9] SSIMLayer: Towards Robust Deep Representation Learning via Nonlinear Structural Similarity
    Abobakr, Ahmed
    Hossny, Mohammed
    Nahavandi, Saeid
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 1234 - 1238
  • [10] Robust Face Recognition via Sparse Representation
    Wright, John
    Yang, Allen Y.
    Ganesh, Arvind
    Sastry, S. Shankar
    Ma, Yi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (02) : 210 - 227