Attention Calibration for Transformer-based Sequential Recommendation

被引:13
|
作者
Zhou, Peilin [1 ]
Ye, Qichen [2 ]
Xie, Yueqi [1 ]
Gao, Jingqi [3 ]
Wang, Shoujin [4 ]
Kim, Jae Boum [1 ]
You, Chenyu [5 ]
Kim, Sunghun [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Guangzhou, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
[3] Upstage, Salt Lake City, UT USA
[4] Univ Technol Sydney, Sydney, NSW, Australia
[5] Yale Univ, New Haven, CT 06520 USA
关键词
Sequential Recommendation; Attention Mechanism; Transformer;
D O I
10.1145/3583780.3614785
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based sequential recommendation (SR) has been booming in recent years, with the self-attention mechanism as its key component. Self-attention has been widely believed to be able to effectively select those informative and relevant items from a sequence of interacted items for next-item prediction via learning larger attention weights for these items. However, this may not always be true in reality. Our empirical analysis of some representative Transformer-based SR models reveals that it is not uncommon for large attention weights to be assigned to less relevant items, which can result in inaccurate recommendations. Through further in-depth analysis, we find two factors that may contribute to such inaccurate assignment of attention weights: sub-optimal position encoding and noisy input. To this end, in this paper, we aim to address this significant yet challenging gap in existing works. To be specific, we propose a simple yet effective framework called Attention Calibration for Transformer-based Sequential Recommendation (AC-TSR). In AC-TSR, a novel spatial calibrator and adversarial calibrator are designed respectively to directly calibrates those incorrectly assigned attention weights. The former is devised to explicitly capture the spatial relationships (i.e., order and distance) among items for more precise calculation of attention weights. The latter aims to redistribute the attention weights based on each item's contribution to the next-item prediction. AC-TSR is readily adaptable and can be seamlessly integrated into various existing transformerbased SR models. Extensive experimental results on four benchmark real-world datasets demonstrate the superiority of our proposed AC-TSR via significant recommendation performance enhancements. The source code is available at https://github.com/AIM- SE/AC-TSR.
引用
收藏
页码:3595 / 3605
页数:11
相关论文
共 50 条
  • [31] Calibration of Transformer-Based Models for Identifying Stress and Depression in Social Media
    Ilias, Loukas
    Mouzakitis, Spiros
    Askounis, Dimitris
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (02) : 1979 - 1990
  • [32] CalibFormer: A Transformer-based Automatic LiDAR-Camera Calibration Network
    Xiao, Yuxuan
    Li, Yao
    Meng, Chengzhen
    Li, Xingchen
    Ji, Jianmin
    Zhang, Yanyong
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 16714 - 16720
  • [33] Improving Conversational Recommender Systems via Transformer-based Sequential Modelling
    Zou, Jie
    Kanoulas, Evangelos
    Ren, Pengjie
    Ren, Zhaochun
    Sun, Aixin
    Long, Cheng
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2319 - 2324
  • [34] Hierarchical Image Generation via Transformer-Based Sequential Patch Selection
    Xu, Xiaogang
    Xu, Ning
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 2938 - 2945
  • [35] Adaptive Disentangled Transformer for Sequential Recommendation
    Zhang, Yipeng
    Wang, Xin
    Chen, Hong
    Zhu, Wenwu
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 3434 - 3445
  • [36] Knowledge Graph Transformer for Sequential Recommendation
    Zhu, Jinghua
    Cui, Yanchang
    Zhang, Zhuohao
    Xi, Heran
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VI, 2023, 14259 : 459 - 471
  • [37] Sequential recommendation by reprogramming pretrained transformer
    Tang, Min
    Cui, Shujie
    Jin, Zhe
    Liang, Shiuan-ni
    Li, Chenliang
    Zou, Lixin
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (01)
  • [38] A Transformer-Based Substitute Recommendation Model IncorporatingWeakly Supervised Customer Behavior Data
    Ye, Wenting
    Yang, Hongfei
    Zhao, Shuai
    Fang, Haoyang
    Shi, Xingjian
    Neppalli, Naveen
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 3325 - 3329
  • [39] Gated Attention with Asymmetric Regularization for Transformer-based Continual Graph Learning
    Lin, Hongxiang
    Jia, Ruiqi
    Lyu, Xiaoqing
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2021 - 2025
  • [40] Transformer-based Scene Graph Generation Network With Relational Attention Module
    Yamamoto, Takuma
    Obinata, Yuya
    Nakayama, Osafumi
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2034 - 2041