MERGE: A Modal Equilibrium Relational Graph Framework for Multi-Modal Knowledge Graph Completion

被引:0
|
作者
Shang, Yuying [1 ,2 ,3 ,4 ]
Fu, Kun [1 ,2 ,3 ]
Zhang, Zequn [1 ,2 ]
Jin, Li [1 ,2 ]
Liu, Zinan [1 ,3 ,4 ]
Wang, Shensi [1 ,2 ,3 ,4 ]
Li, Shuchao [1 ,2 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Beijing 100094, Peoples R China
[2] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Network Informat Syst Technol NIST, Beijing 100190, Peoples R China
[3] Univ Chinese Acad Sci, Beijing 100190, Peoples R China
[4] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100094, Peoples R China
基金
中国国家自然科学基金;
关键词
multi-modal knowledge graph; knowledge graph representation; graph attention network; information integration;
D O I
10.3390/s24237605
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The multi-modal knowledge graph completion (MMKGC) task aims to automatically mine the missing factual knowledge from the existing multi-modal knowledge graphs (MMKGs), which is crucial in advancing cross-modal learning and reasoning. However, few methods consider the adverse effects caused by different missing modal information in the model learning process. To address the above challenges, we innovatively propose a Modal Equilibrium Relational Graph framEwork, called MERGE. By constructing three modal-specific directed relational graph attention networks, MERGE can implicitly represent missing modal information for entities by aggregating the modal embeddings from neighboring nodes. Subsequently, a fusion approach based on low-rank tensor decomposition is adopted to align multiple modal features in both the explicit structural level and the implicit semantic level, utilizing the structural information inherent in the original knowledge graphs, which enhances the interpretability of the fused features. Furthermore, we introduce a novel interpolation re-ranking strategy to adjust the importance of modalities during inference while preserving the semantic integrity of each modality. The proposed framework has been validated on four publicly available datasets, and the experimental results have demonstrated the effectiveness and robustness of our method in the MMKGC task.
引用
收藏
页数:30
相关论文
共 50 条
  • [1] Hyper-node Relational Graph Attention Network for Multi-modal Knowledge Graph Completion
    Liang, Shuang
    Zhu, Anjie
    Zhang, Jiasheng
    Shao, Jie
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (02)
  • [2] NativE: Multi-modal Knowledge Graph Completion in the Wild
    Zhang, Yichi
    Chen, Zhuo
    Guo, Lingbing
    Xu, Yajing
    Hu, Binbin
    Liu, Ziqi
    Zhang, Wen
    Chen, Huajun
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 91 - 101
  • [3] Multi-Modal Knowledge Graph Transformer Framework for Multi-Modal Entity Alignment
    Li, Qian
    Ji, Cheng
    Guo, Shu
    Liang, Zhaoji
    Wang, Lihong
    Li, Jianxin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 987 - 999
  • [4] MultiJAF: Multi-modal joint entity alignment framework for multi-modal knowledge graph
    Cheng, Bo
    Zhu, Jia
    Guo, Meimei
    NEUROCOMPUTING, 2022, 500 : 581 - 591
  • [5] Richpedia: A Comprehensive Multi-modal Knowledge Graph
    Wang, Meng
    Qi, Guilin
    Wang, Haofen
    Zheng, Qiushuo
    SEMANTIC TECHNOLOGY, JIST 2019: PROCEEDINGS, 2020, 12032 : 130 - 145
  • [6] What Is a Multi-Modal Knowledge Graph: A Survey
    Peng, Jinghui
    Hu, Xinyu
    Huang, Wenbo
    Yang, Jian
    BIG DATA RESEARCH, 2023, 32
  • [7] Multi-Modal Relational Graph for Cross-Modal Video Moment Retrieval
    Zeng, Yawen
    Cao, Da
    Wei, Xiaochi
    Liu, Meng
    Zhao, Zhou
    Qin, Zheng
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 2215 - 2224
  • [8] Multi-modal Graph Convolutional Network for Knowledge Graph Entity Alignment
    You, Yinghui
    Wei, Yuyang
    Zhang, Yanlong
    Chen, Wei
    Zhao, Lei
    WEB AND BIG DATA, PT I, APWEB-WAIM 2023, 2024, 14331 : 142 - 157
  • [9] MLSFF: Multi-level structural features fusion for multi-modal knowledge graph completion
    Zhai, Hanming
    Lv, Xiaojun
    Hou, Zhiwen
    Tong, Xin
    Bu, Fanliang
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (08) : 14096 - 14116
  • [10] Contrastive Multi-Modal Knowledge Graph Representation Learning
    Fang, Quan
    Zhang, Xiaowei
    Hu, Jun
    Wu, Xian
    Xu, Changsheng
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (09) : 8983 - 8996