Sequential recommendation by reprogramming pretrained transformer

被引:0
|
作者
Tang, Min [1 ]
Cui, Shujie [2 ]
Jin, Zhe [3 ]
Liang, Shiuan-ni [1 ]
Li, Chenliang [4 ]
Zou, Lixin [4 ]
机构
[1] Monash Univ, Sch Engn, Bandar Sunway 47500, Malaysia
[2] Monash Univ, Sch Informat Technol, Clayton, Vic 3800, Australia
[3] Anhui Univ, Sch Artificial Intelligence, Hefei 230039, Anhui, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Generative pretrained transformer; Few-shot learning;
D O I
10.1016/j.ipm.2024.103938
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inspired by the success of Pre-trained language models (PLMs), numerous sequential recommenders attempted to replicate its achievements by employing PLMs' efficient architectures for building large models and using self-supervised learning for broadening training data. Despite their success, there is curiosity about developing a large-scale sequential recommender system since existing methods either build models within a single dataset or utilize text as an intermediary for alignment across different datasets. However, due to the sparsity of user- item interactions, unalignment between different datasets, and lack of global information in the sequential recommendation, directly pre-training a large foundation model may not be feasible. Towards this end, we propose the RecPPT that firstly employs the GPT-2 to model historical sequence by training the input item embedding and the output layer from scratch, which avoids training a large model on the sparse user-item interactions. Additionally, to alleviate the burden of unalignment, the RecPPT is equipped with a reprogramming module to reprogram the target embedding to existing well-trained proto-embeddings. Furthermore, RecPPT integrates global information into sequences by initializing the item embedding using an SVD-based initializer. Extensive experiments over four datasets demonstrated the RecPPT achieved an average improvement of 6.5% on NDCG@5, 6.2% on NDCG@10, 6.1% on Recall@5, and 5.4% on Recall@10 compared to the baselines. Particularly in few-shot scenarios, the significant improvements in NDCG@10 confirm the superiority of the proposed method.
引用
收藏
页数:15
相关论文
共 50 条
  • [11] SSE-PT: Sequential Recommendation Via Personalized Transformer
    Wu, Liwei
    Li, Shuqing
    Hsieh, Cho-Jui
    Sharpnack, James
    RECSYS 2020: 14TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, 2020, : 328 - 337
  • [12] Transformer-Based Rating-Aware Sequential Recommendation
    Li, Yang
    Li, Qianmu
    Meng, Shunmei
    Hou, Jun
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT I, 2022, 13155 : 759 - 774
  • [13] Dual Contrastive Transformer for Hierarchical Preference Modeling in Sequential Recommendation
    Huang, Chengkai
    Wang, Shoujin
    Wang, Xianzhi
    Yao, Lina
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 99 - 109
  • [14] DGT: Unbiased sequential recommendation via Disentangled Graph Transformer
    Li, Chenglin
    Xie, Tao
    Yu, Chenyun
    Hu, Bo
    Li, Zang
    Cheng, Lei
    Kong, Beibei
    Niu, Di
    KNOWLEDGE-BASED SYSTEMS, 2025, 310
  • [15] Time-Aware Squeeze-Excitation Transformer for Sequential Recommendation
    Chen, Hongwei
    Liu, Luanxuan
    Chen, Zexi
    Li, Xia
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT IX, 2024, 15024 : 121 - 135
  • [16] Multi-scale Interest Dynamic Hierarchical Transformer for sequential recommendation
    Nana Huang
    Ruimin Hu
    Mingfu Xiong
    Xiaoran Peng
    Hongwei Ding
    Xiaodong Jia
    Lingkun Zhang
    Neural Computing and Applications, 2022, 34 : 16643 - 16654
  • [17] Personalization Through User Attributes for Transformer-Based Sequential Recommendation
    Fischer, Elisabeth
    Dallmann, Alexander
    Hotho, Andreas
    RECOMMENDER SYSTEMS IN FASHION AND RETAIL, 2023, 981 : 25 - 43
  • [18] Continuous-Time Sequential Recommendation with Temporal Graph Collaborative Transformer
    Fan, Ziwei
    Liu, Zhiwei
    Zhang, Jiawei
    Xiong, Yun
    Zheng, Lei
    Yu, Philip S.
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 433 - 442
  • [19] Multi-scale Interest Dynamic Hierarchical Transformer for sequential recommendation
    Huang, Nana
    Hu, Ruimin
    Xiong, Mingfu
    Peng, Xiaoran
    Ding, Hongwei
    Jia, Xiaodong
    Zhang, Lingkun
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (19): : 16643 - 16654
  • [20] A global contextual enhanced structural-aware transformer for sequential recommendation
    Zhang, Zhu
    Yang, Bo
    Chen, Xingming
    Li, Qing
    KNOWLEDGE-BASED SYSTEMS, 2024, 304