Multi-modal entity alignment based on joint knowledge representation learning

被引:0
|
作者
Wang H.-Y. [1 ]
Lun B. [1 ]
Zhang X.-M. [1 ]
Sun X.-L. [1 ]
机构
[1] College of Information Science and Engineering, Heibei University of Science and Technology, Shijiazhuang
来源
Kongzhi yu Juece/Control and Decision | 2021年 / 35卷 / 12期
关键词
Knowledge graphs (KGs); Knowledge representation learning (KRL); Multi-modal data; Multi-modal entity alignment; Seed set; Translating model;
D O I
10.13195/j.kzyjc.2019.0331
中图分类号
学科分类号
摘要
Knowledge representation learning (KRL) based entity alignment represents multi-knowledge graphs (KGs) as low-dimensional embeddings and realizes entity alignment by measuring the similarities between entity vectors. Existing approaches, however, frequently primarily focus on the text information but ignore the image information, and lead entity feature information of the image to being underutilized. To address this problem, we propose an approach for image-text multi-modal entity alignment (ITMEA) via joint knowledge representation learning. The approach joints multi-modal (image and text) data, and embeds multi-model data into a uniform low-dimensional semantic space by a knowledge representation learning model which combines the translating embedding model (TransE) with the dynamic mapping matrix embedding model (TransD). In low-dimensional semantic space, the link-mapping relations among the aligned multi-modal entities can be learned iteratively from the seed set. The learned link-mapping relations can be applied to unaligned entities. In this way, the multi-modal entity alignment can be implemented. Experiment results on WN18-IMG datasets show that the ITMEA can achieve the multi-modal entity alignment, and it is useful for the construction and completion of the open knowledge graph. Copyright ©2020 Control and Decision.
引用
收藏
页码:2855 / 2864
页数:9
相关论文
共 38 条
  • [21] Zhu H, Xie R B, Liu Z Y, Et al., Iterative entity alignment via joint knowledge embeddings, Proceedings of the 26th International Joint Conference on Artificial Intelligence, pp. 4258-4264, (2017)
  • [22] Pershina M, Yakout M, Chakrabarti K., Holistic entity matching across knowledge graphs, Proceedings of the IEEE International Conference on Cloud and Big Data Computing, pp. 1585-1590, (2015)
  • [23] Sun Z Q, Hu W, Li C K., Cross-lingual entity alignment via joint attribute-preserving embedding, Proceedings of the 16th International Semantic Web Conference, pp. 628-644, (2017)
  • [24] Socher R, Chen D, Manning C D, Et al., Reasoning with neural tensor networks for knowledge base completion, Proceedings of the International Conference on Neural Information Processing Systems, pp. 926-934, (2013)
  • [25] Xiao H, Huang M, Meng L, Et al., SSP: Semantic space projection for knowledge graph embedding with text descriptions, Proceedings of the 31st Association for the Advancement of Artificial Intelligence, pp. 3104-3110, (2017)
  • [26] Wang Z, Zhang J, Feng J, Et al., Knowledge graph embedding by translating on hyperplanes, Proceedings of the 28th Association for the Advancement of Artificial Intelligence, pp. 1112-1119, (2014)
  • [27] Lin Y, Liu Z, Zhu X, Et al., Learning entity and relation embeddings for knowledge graph completion, Proceedings of the 29th Association for the Advancement of Artificial Intelligence, pp. 2181-2187, (2015)
  • [28] Dettmers T, Minervini P, Stenetorp P, Et al., Convolutional 2d knowledge graph embeddings, Proceedings of the 32nd Association for the Advancement of Artificial Intelligence, pp. 1811-1818, (2018)
  • [29] Nguyen D Q, Nguyen T D, Nguyen D Q, Et al., A novel embedding model for knowledge base completion based on convolutional neural network, Proceedings of the 16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 327-333, (2018)
  • [30] Pezeshkpour P, Chen L Y, Singh S., Embedding multimodal relational data for knowledge base completion, Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 3208-3218, (2018)