Multi-modal entity alignment based on joint knowledge representation learning

被引:0
|
作者
Wang H.-Y. [1 ]
Lun B. [1 ]
Zhang X.-M. [1 ]
Sun X.-L. [1 ]
机构
[1] College of Information Science and Engineering, Heibei University of Science and Technology, Shijiazhuang
来源
Kongzhi yu Juece/Control and Decision | 2021年 / 35卷 / 12期
关键词
Knowledge graphs (KGs); Knowledge representation learning (KRL); Multi-modal data; Multi-modal entity alignment; Seed set; Translating model;
D O I
10.13195/j.kzyjc.2019.0331
中图分类号
学科分类号
摘要
Knowledge representation learning (KRL) based entity alignment represents multi-knowledge graphs (KGs) as low-dimensional embeddings and realizes entity alignment by measuring the similarities between entity vectors. Existing approaches, however, frequently primarily focus on the text information but ignore the image information, and lead entity feature information of the image to being underutilized. To address this problem, we propose an approach for image-text multi-modal entity alignment (ITMEA) via joint knowledge representation learning. The approach joints multi-modal (image and text) data, and embeds multi-model data into a uniform low-dimensional semantic space by a knowledge representation learning model which combines the translating embedding model (TransE) with the dynamic mapping matrix embedding model (TransD). In low-dimensional semantic space, the link-mapping relations among the aligned multi-modal entities can be learned iteratively from the seed set. The learned link-mapping relations can be applied to unaligned entities. In this way, the multi-modal entity alignment can be implemented. Experiment results on WN18-IMG datasets show that the ITMEA can achieve the multi-modal entity alignment, and it is useful for the construction and completion of the open knowledge graph. Copyright ©2020 Control and Decision.
引用
收藏
页码:2855 / 2864
页数:9
相关论文
共 38 条
  • [1] Yang F, Yang Z L, Cohen W W., Differentiable learning of logical rules for knowledge base reasoning, Proceedings of the International Conference on Neural Information Processing Systems, pp. 2319-2328, (2017)
  • [2] Dubey M, Banerjee D, Chaudhuri D, Et al., EARL: Joint entity and relation linking for question answering over knowledge graphs, Proceedings of the 17th International Semantic Web Conference, pp. 108-126, (2018)
  • [3] Wang H W, Zhang F Z, Xie X, Et al., DKN: Deep knowledge-aware network for new recommendation, Proceedings of the 2018 World Wide Web Conference, pp. 1835-1844, (2018)
  • [4] Suchanek F M, Kasneci G, Weikum G., Yago: A core of semantic knowledge, Proceedings of the 2007 International world Wide Web Conference, pp. 697-706, (2007)
  • [5] Lehmann J, Isele R, Jakob M, Et al., DBpedia — A large-scale, multilingual knowledge base extracted from wikipedia, Semantic Web, 6, 2, pp. 167-195, (2015)
  • [6] Bollacker K, Evans C, Paritosh P, Et al., Freebase: A collaboratively created graph database for structuring human knowledge, Proceedings of the 14th ACM SIGMOD International Conference on Management of data, pp. 1247-1250, (2008)
  • [7] Liu Z Y, Sun M S, Lin Y K, Et al., Knowledge representation learning: A review, Journal of Computer Research and Development, 53, 2, pp. 247-261, (2016)
  • [8] Bordes A, Usunier N, Garcia-Duran A, Et al., Translating embeddings for modeling multi-relational data, Proceedings of the International Conference on Neural Information Processing Systems, pp. 2787-2795, (2013)
  • [9] Antol S, Agrawal A, Lu J, Et al., VQA: Visual question answering, Proceedings of the IEEE International Conference on Computer Vision, pp. 2425-2433, (2015)
  • [10] Huang F Q, Zhang X M, Xu J, Et al., Network embedding by fusing multimodal contents and links, Knowledge-Based Systems, 171, pp. 44-55, (2019)