Attention Calibration for Transformer-based Sequential Recommendation

被引:13
|
作者
Zhou, Peilin [1 ]
Ye, Qichen [2 ]
Xie, Yueqi [1 ]
Gao, Jingqi [3 ]
Wang, Shoujin [4 ]
Kim, Jae Boum [1 ]
You, Chenyu [5 ]
Kim, Sunghun [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Guangzhou, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
[3] Upstage, Salt Lake City, UT USA
[4] Univ Technol Sydney, Sydney, NSW, Australia
[5] Yale Univ, New Haven, CT 06520 USA
关键词
Sequential Recommendation; Attention Mechanism; Transformer;
D O I
10.1145/3583780.3614785
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based sequential recommendation (SR) has been booming in recent years, with the self-attention mechanism as its key component. Self-attention has been widely believed to be able to effectively select those informative and relevant items from a sequence of interacted items for next-item prediction via learning larger attention weights for these items. However, this may not always be true in reality. Our empirical analysis of some representative Transformer-based SR models reveals that it is not uncommon for large attention weights to be assigned to less relevant items, which can result in inaccurate recommendations. Through further in-depth analysis, we find two factors that may contribute to such inaccurate assignment of attention weights: sub-optimal position encoding and noisy input. To this end, in this paper, we aim to address this significant yet challenging gap in existing works. To be specific, we propose a simple yet effective framework called Attention Calibration for Transformer-based Sequential Recommendation (AC-TSR). In AC-TSR, a novel spatial calibrator and adversarial calibrator are designed respectively to directly calibrates those incorrectly assigned attention weights. The former is devised to explicitly capture the spatial relationships (i.e., order and distance) among items for more precise calculation of attention weights. The latter aims to redistribute the attention weights based on each item's contribution to the next-item prediction. AC-TSR is readily adaptable and can be seamlessly integrated into various existing transformerbased SR models. Extensive experimental results on four benchmark real-world datasets demonstrate the superiority of our proposed AC-TSR via significant recommendation performance enhancements. The source code is available at https://github.com/AIM- SE/AC-TSR.
引用
收藏
页码:3595 / 3605
页数:11
相关论文
共 50 条
  • [41] TRANSFORMER-BASED LIP-READING WITH REGULARIZED DROPOUT AND RELAXED ATTENTION
    Li, Zhengyang
    Lohrenz, Timo
    Dunkelberg, Matthias
    Fingscheidt, Tim
    2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 723 - 730
  • [42] Transformer-Based Attention Networks for Continuous Pixel-Wise Prediction
    Yang, Guanglei
    Tang, Hao
    Ding, Mingli
    Sebe, Nicu
    Ricci, Elisa
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 16249 - 16259
  • [43] Roles and Utilization of Attention Heads in Transformer-based Neural Language Models
    Jo, Jae-young
    Myaeng, Sung-hyon
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3404 - 3417
  • [44] Adaptive Sparse and Monotonic Attention for Transformer-based Automatic Speech Recognition
    Zhao, Chendong
    Wang, Jianzong
    Wei, Wenqi
    Qu, Xiaoyang
    Wang, Haoqian
    Xiao, Jing
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2022, : 173 - 180
  • [45] A hierarchical contextual attention-based network for sequential recommendation
    Cui, Qiang
    Wu, Shu
    Huang, Yan
    Wang, Liang
    NEUROCOMPUTING, 2019, 358 : 141 - 149
  • [46] Sequential Recommendation Based on Multivariate Hawkes Process Embedding With Attention
    Wang, Dongjing
    Zhang, Xin
    Xiang, Zhengzhe
    Yu, Dongjin
    Xu, Guandong
    Deng, Shuiguang
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (11) : 11893 - 11905
  • [47] TransDARC: Transformer-based Driver Activity Recognition with Latent Space Feature Calibration
    Peng, Kunyu
    Roitberg, Alina
    Yang, Kailun
    Zhang, Jiaming
    Stiefelhagen, Rainer
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 278 - 285
  • [48] Improving Transformer-based Sequential Conversational Recommendations through Knowledge Graph Embeddings
    Petruzzelli, Alessandro
    Martina, Alessandro Francesco Maria
    Spillo, Giuseppe
    Musto, Cataldo
    de Gemmis, Marco
    Lops, Pasquale
    Semeraro, Giovanni
    PROCEEDINGS OF THE 32ND ACM CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION, UMAP 2024, 2024, : 172 - 182
  • [49] G-TransRec: A Transformer-Based Next-Item Recommendation With Time Prediction
    Chen, Yi-Cheng
    Chen, Yen-Liang
    Hsu, Chia-Hsiang
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (03) : 4175 - 4188
  • [50] TransDTI: Transformer-Based Language Models for Estimating DTIs and Building a Drug Recommendation Workflow
    Kalakoti, Yogesh
    Yadav, Shashank
    Sundar, Durai
    ACS OMEGA, 2022, 7 (03): : 2706 - 2717