Guiding Graph Learning with Denoised Modality for Multi-modal Recommendation

被引:0
|
作者
Wang, Yuexian [1 ]
Ma, Wenze [1 ]
Zhu, Yanmin [1 ]
Wang, Chunyang [1 ]
Wang, Zhaobo [1 ]
Tang, Feilong [1 ]
Yu, Jiadi [1 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai 201109, Peoples R China
基金
美国国家科学基金会;
关键词
Multi-modal Recommendation; Graph Structure Learning; Self-supervised Learning;
D O I
10.1007/978-981-97-5572-1_14
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-modal recommendation improves the recommendation accuracy by leveraging various modalities (e.g., visual, textual, and acoustic) of rich item content. However, most existing studies overlook that modality features can be noisy for recommendation. Recently, several graph-based methods have attempted to alleviate the noise issue via graph structure learning, but they roughly focus on edge denoising while neglecting node denoising. This will lead to the following circumstances: 1) noisy factors at the node level and 2) insufficient edge-level denoising. To address the limitations, we propose a Denoised Modality-guided Graph Learning paradigm (DMGL), which could jointly and iteratively eliminate both the node-level and edge-level noise for multi-modal recommendation. Meanwhile, masked feature autoencoder and contrastive learning mechanism are introduced to handle intra- and inter-modality node-level noise, respectively. Extensive experiments on real-world datasets demonstrate the superior performance of our proposed model. The codes are available here.
引用
收藏
页码:220 / 235
页数:16
相关论文
共 50 条
  • [21] Multi-Modal Transportation Recommendation with Unified Route Representation Learning
    Liu, Hao
    Han, Jindong
    Fu, Yanjie
    Zhou, Jingbo
    Lu, Xinjiang
    Xiong, Hui
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2020, 14 (03): : 342 - 350
  • [22] UaMC: user-augmented conversation recommendation via multi-modal graph learning and context mining
    Fan, Siqi
    Wang, Yequan
    Pang, Xiaobing
    Chen, Lisi
    Han, Peng
    Shang, Shuo
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2023, 26 (06): : 4109 - 4129
  • [23] MULTI-MODAL REPRESENTATION LEARNING FOR SHORT VIDEO UNDERSTANDING AND RECOMMENDATION
    Guo, Daya
    Hong, Jiangshui
    Luo, Binli
    Yan, Qirui
    Niu, Zhangming
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 687 - 690
  • [24] UaMC: user-augmented conversation recommendation via multi-modal graph learning and context mining
    Siqi Fan
    Yequan Wang
    Xiaobing Pang
    Lisi Chen
    Peng Han
    Shuo Shang
    World Wide Web, 2023, 26 : 4109 - 4129
  • [25] Cross-view Contrastive Learning Enhanced Heterogeneous Graph Networks for Multi-modal Recipe Recommendation
    She, Xiaolong
    Zhou, Jingya
    Hu, Zhenyu
    Du, Boyu
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2024, PT 3, 2025, 14852 : 467 - 482
  • [26] Modality-Aware Negative Sampling for Multi-modal Knowledge Graph Embedding
    Zhang, Yichi
    Chen, Mingyang
    Zhang, Wen
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [27] An Enhanced Multi-Modal Recommendation Based on Alternate Training With Knowledge Graph Representation
    Wang, Yuequn
    Dong, Liyan
    Zhang, Hao
    Ma, Xintao
    Li, Yongli
    Sun, Minghui
    IEEE ACCESS, 2020, 8 : 213012 - 213026
  • [28] Learnable Cross-modal Knowledge Distillation for Multi-modal Learning with Missing Modality
    Wang, Hu
    Ma, Congbo
    Zhang, Jianpeng
    Zhang, Yuan
    Avery, Jodie
    Hull, Louise
    Carneiro, Gustavo
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT IV, 2023, 14223 : 216 - 226
  • [29] Overcome Modal Bias in Multi-modal Federated Learning via Balanced Modality Selection
    Fan, Yunfeng
    Xu, Wenchao
    Wang, Haozhao
    Huo, Fushuo
    Chen, Jinyu
    Guo, Song
    COMPUTER VISION - ECCV 2024, PT LXXXI, 2025, 15139 : 178 - 195
  • [30] Multi-Modal Correction Network for Recommendation
    Wang, Zengmao
    Feng, Yunzhen
    Zhang, Xin
    Yang, Renjie
    Du, Bo
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (02) : 810 - 822