Sequential recommendation by reprogramming pretrained transformer

被引:0
|
作者
Tang, Min [1 ]
Cui, Shujie [2 ]
Jin, Zhe [3 ]
Liang, Shiuan-ni [1 ]
Li, Chenliang [4 ]
Zou, Lixin [4 ]
机构
[1] Monash Univ, Sch Engn, Bandar Sunway 47500, Malaysia
[2] Monash Univ, Sch Informat Technol, Clayton, Vic 3800, Australia
[3] Anhui Univ, Sch Artificial Intelligence, Hefei 230039, Anhui, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Generative pretrained transformer; Few-shot learning;
D O I
10.1016/j.ipm.2024.103938
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inspired by the success of Pre-trained language models (PLMs), numerous sequential recommenders attempted to replicate its achievements by employing PLMs' efficient architectures for building large models and using self-supervised learning for broadening training data. Despite their success, there is curiosity about developing a large-scale sequential recommender system since existing methods either build models within a single dataset or utilize text as an intermediary for alignment across different datasets. However, due to the sparsity of user- item interactions, unalignment between different datasets, and lack of global information in the sequential recommendation, directly pre-training a large foundation model may not be feasible. Towards this end, we propose the RecPPT that firstly employs the GPT-2 to model historical sequence by training the input item embedding and the output layer from scratch, which avoids training a large model on the sparse user-item interactions. Additionally, to alleviate the burden of unalignment, the RecPPT is equipped with a reprogramming module to reprogram the target embedding to existing well-trained proto-embeddings. Furthermore, RecPPT integrates global information into sequences by initializing the item embedding using an SVD-based initializer. Extensive experiments over four datasets demonstrated the RecPPT achieved an average improvement of 6.5% on NDCG@5, 6.2% on NDCG@10, 6.1% on Recall@5, and 5.4% on Recall@10 compared to the baselines. Particularly in few-shot scenarios, the significant improvements in NDCG@10 confirm the superiority of the proposed method.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Evaluating the Effectiveness of a Generative Pretrained Transformer-Based Dietary Recommendation System in Managing Potassium Intake for Hemodialysis Patients
    Jin, Haijiao
    Lin, Qisheng
    Lu, Jifang
    Hu, Cuirong
    Lu, Bohan
    Jiang, Na
    Wu, Shaun
    Li, Xiaoyang
    JOURNAL OF RENAL NUTRITION, 2024, 34 (06) : 539 - 545
  • [32] A Heideggerian analysis of generative pretrained transformer models
    Floroiu, Iustin
    Timisica, Daniela
    ROMANIAN JOURNAL OF INFORMATION TECHNOLOGY AND AUTOMATIC CONTROL-REVISTA ROMANA DE INFORMATICA SI AUTOMATICA, 2024, 34 (01): : 13 - 22
  • [33] Knowledge-Enhanced Conversational Recommendation via Transformer-Based Sequential Modeling
    Zou, Jie
    Sun, Aixin
    Long, Cheng
    Kanoulas, Evangelos
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (06)
  • [34] BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer
    Sun, Fei
    Liu, Jun
    Wu, Jian
    Pei, Changhua
    Lin, Xiao
    Ou, Wenwu
    Jiang, Peng
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 1441 - 1450
  • [35] Pretrained Language Models for Sequential Sentence Classification
    Cohan, Arman
    Beltagy, Iz
    King, Daniel
    Dalvi, Bhavana
    Weld, Daniel S.
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3693 - 3699
  • [36] Graph Transformer for Recommendation
    Li, Chaoliu
    Xia, Lianghao
    Ren, Xubin
    Ye, Yaowen
    Xu, Yong
    Huang, Chao
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 1680 - 1689
  • [37] ADAPTING PRETRAINED TRANSFORMER TO LATTICES FOR SPOKEN LANGUAGE UNDERSTANDING
    Huang, Chao-Wei
    Chen, Yun-Nung
    2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 845 - 852
  • [38] Is News Recommendation a Sequential Recommendation Task?
    Wu, Chuhan
    Wu, Fangzhao
    Qi, Tao
    Li, Chenliang
    Huang, Yongfeng
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2382 - 2386
  • [39] Chat generative pretrained transformer: A disruptive or constructive technology?
    Deshmukh, Sonali Vijay
    JOURNAL OF THE INTERNATIONAL CLINICAL DENTAL RESEARCH ORGANIZATION, 2023, 15 (01) : 1 - 2
  • [40] Extracting Sentence Embeddings from Pretrained Transformer Models
    Stankevicius, Lukas
    Lukosevicius, Mantas
    APPLIED SCIENCES-BASEL, 2024, 14 (19):