Personalized Showcases: Generating Multi-Modal Explanations for Recommendations

被引:4
|
作者
Yan, An [1 ]
He, Zhankui [1 ]
Li, Jiacheng [1 ]
Zhang, Tianyang [1 ]
McAuley, Julian [1 ]
机构
[1] Univ Calif San Diego, La Jolla, CA USA
来源
PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023 | 2023年
关键词
Datasets; Text Generation; Multi-Modality; Contrastive Learning;
D O I
10.1145/3539618.3592036
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing explanation models generate only text for recommendations but still struggle to produce diverse contents. In this paper, to further enrich explanations, we propose a new task named personalized showcases, in which we provide both textual and visual information to explain our recommendations. Specifically, we first select a personalized image set that is the most relevant to a user's interest toward a recommended item. Then, natural language explanations are generated accordingly given our selected images. For this new task, we collect a large-scale dataset from Google Maps and construct a high-quality subset for generating multi-modal explanations. We propose a personalized multi-modal framework which can generate diverse and visually-aligned explanations via contrastive learning. Experiments show that our framework benefits from different modalities as inputs, and is able to produce more diverse and expressive explanations compared to previous methods on a variety of evaluation metrics. (1)
引用
收藏
页码:2251 / 2255
页数:5
相关论文
共 50 条
  • [41] Conversational multi-modal browser: An integrated multi-modal browser and dialog manager
    Tiwari, A
    Hosn, RA
    Maes, SH
    2003 SYMPOSIUM ON APPLICATIONS AND THE INTERNET, PROCEEDINGS, 2003, : 348 - 351
  • [42] Hierarchical Multi-Modal Prompting Transformer for Multi-Modal Long Document Classification
    Liu, Tengfei
    Hu, Yongli
    Gao, Junbin
    Sun, Yanfeng
    Yin, Baocai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (07) : 6376 - 6390
  • [43] Generating and Understanding Personalized Explanations in Hybrid Recommender Systems
    Kouki, Pigi
    Schaffer, James
    Pujara, Jay
    O'Donovan, John
    Getoor, Lise
    ACM TRANSACTIONS ON INTERACTIVE INTELLIGENT SYSTEMS, 2020, 10 (04)
  • [44] Session-based news recommendations using SimRank on multi-modal graphs
    Symeonidis, Panagiotis
    Kirjackaja, Lidija
    Zanker, Markus
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 180
  • [45] A Comprehensive Multi-modal Approach for Enhanced Product Recommendations Based on Customer Habits
    Bodapati J.D.
    Veeranjaneyulu N.
    Yenduri L.K.
    Journal of The Institution of Engineers (India): Series B, 2024, 105 (06) : 1537 - 1545
  • [46] Visual Explanations of Image-Text Representations via Multi-Modal Information Bottleneck Attribution
    Wang, Ying
    Rudner, Tim G. J.
    Wilson, Andrew Gordon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [47] Multi-modal pedestrian detection with misalignment based on modal-wise regression and multi-modal IoU
    Wanchaitanawong, Napat
    Tanaka, Masayuki
    Shibata, Takashi
    Okutomi, Masatoshi
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (01)
  • [48] MCCP: multi-modal fashion compatibility and conditional preference model for personalized clothing recommendation
    Wang, Yunzhu
    Liu, Li
    Fu, Xiaodong
    Liu, Lijun
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (04) : 9621 - 9645
  • [49] Personalized retrieval of sports video based on multi-modal analysis and user preference acquisition
    Zhang, Yi-Fan
    Xu, Changsheng
    Zhang, Xiaoyu
    Lu, Hanqing
    MULTIMEDIA TOOLS AND APPLICATIONS, 2009, 44 (02) : 305 - 330
  • [50] Multi-Modal LA in Personalized Education Using Deep Reinforcement Learning Based Approach
    Sharif, Muddsair
    Uckelmann, Dieter
    IEEE ACCESS, 2024, 12 : 54049 - 54065