VEG-MMKG: Multimodal knowledge graph construction for vegetables based on pre-trained model extraction

被引:0
|
作者
Lv, Bowen [1 ,2 ,3 ,4 ]
Wu, Huarui [1 ,3 ,4 ]
Chen, Wenbai [2 ]
Chen, Cheng [1 ]
Miao, Yisheng [1 ,3 ,4 ]
Zhao, Chunjiang [1 ]
机构
[1] Natl Engn Res Ctr Informat Technol Agr, Beijing 100097, Peoples R China
[2] Beijing Informat Sci & Technol Univ, Sch Automat, Beijing 100192, Peoples R China
[3] Beijing Acad Agr & Forestry Sci, Informat Technol Res Ctr, Beijing 100097, Peoples R China
[4] Minist Agr & Rural Affairs, Key Lab Digital Village Technol, Beijing 100097, Peoples R China
关键词
Knowledge graph; Multimodal fusion; Image-text pairs; Pre-trained model;
D O I
10.1016/j.compag.2024.109398
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
Knowledge graph technology is of great significance to modern agricultural information management and datadriven decision support. However, agricultural knowledge is rich in types, and agricultural knowledge graph databases built only based on text are not conducive to users' intuitive perception and comprehensive understanding of knowledge. In view of this, this paper proposes a solution to extract knowledge and construct an agricultural multimodal knowledge graph using a pre-trained language model. This paper takes two plants, cabbage and corn, as research objects. First, a text-image collaborative representation learning method with a two-stream structure is adopted to combine the image modal information of vegetables with the text modal information, and the correlation and complementarity between the two types of information are used to achieve entity alignment. In addition, in order to solve the problem of high similarity of vegetable entities in small categories, a cross-modal fine-grained contrastive learning method is introduced, and the problem of insufficient semantic association between modalities is solved by contrastive learning of vocabulary and small areas of images. Finally, a visual multimodal knowledge graph user interface is constructed using the results of image and text matching. Experimental results show that the image and text matching efficiency of the fine-tuned pretrained model on the vegetable dataset is 76.7%, and appropriate images can be matched for text entities. The constructed visual multimodal knowledge graph database allows users to query and filter knowledge according to their needs, providing assistance for subsequent research on various applications in specific fields such as multimodal agricultural intelligent question and answer, crop pest and disease identification, and agricultural product recommendations.
引用
收藏
页数:13
相关论文
共 50 条
  • [11] Knowledge graph extension with a pre-trained language model via unified learning method
    Choi, Bonggeun
    Ko, Youngjoong
    KNOWLEDGE-BASED SYSTEMS, 2023, 262
  • [12] Fusing graph structural information with pre-trained generative model for knowledge graph-to-text generation
    Shi, Xiayang
    Xia, Zhenlin
    Li, Yinlin
    Wang, Xuhui
    Niu, Yufeng
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (03) : 2619 - 2640
  • [13] SeisCLIP: A Seismology Foundation Model Pre-Trained by Multimodal Data for Multipurpose Seismic Feature Extraction
    Si, Xu
    Wu, Xinming
    Sheng, Hanlin
    Zhu, Jun
    Li, Zefeng
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 13
  • [14] A Pre-trained Knowledge Tracing Model with Limited Data
    Yue, Wenli
    Su, Wei
    Liu, Lei
    Cai, Chuan
    Yuan, Yongna
    Jia, Zhongfeng
    Liu, Jiamin
    Xie, Wenjian
    DATABASE AND EXPERT SYSTEMS APPLICATIONS, PT I, DEXA 2024, 2024, 14910 : 163 - 178
  • [15] MEM-KGC: Masked Entity Model for Knowledge Graph Completion With Pre-Trained Language Model
    Choi, Bonggeun
    Jang, Daesik
    Ko, Youngjoong
    IEEE ACCESS, 2021, 9 : 132025 - 132032
  • [16] Commonsense Knowledge Base Completion with Relational Graph Attention Network and Pre-trained Language Model
    Ju, Jinghao
    Yang, Deqing
    Liu, Jingping
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 4104 - 4108
  • [17] Billion-scale Pre-trained E-commerce Product Knowledge Graph Model
    Zhang, Wen
    Wong, Chi-Man
    Ye, Ganqiang
    Wen, Bo
    Zhang, Wei
    Chen, Huajun
    2021 IEEE 37TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2021), 2021, : 2476 - 2487
  • [18] Enhancing Code Summarization with Graph Embedding and Pre-trained Model
    Li, Lixuan
    Li, Jie
    Xu, Yihui
    Zhu, Hao
    Zhang, Xiaofang
    INTERNATIONAL JOURNAL OF SOFTWARE ENGINEERING AND KNOWLEDGE ENGINEERING, 2023, 33 (11N12) : 1765 - 1786
  • [19] A survey of transformer-based multimodal pre-trained modals
    Han, Xue
    Wang, Yi-Tong
    Feng, Jun-Lan
    Deng, Chao
    Chen, Zhan-Heng
    Huang, Yu-An
    Su, Hui
    Hu, Lun
    Hu, Peng-Wei
    NEUROCOMPUTING, 2023, 515 : 89 - 106
  • [20] Relation extraction method based on pre-trained model and bidirectional semantic union
    He, Xinyu
    Yan, Ge
    Han, Xue
    Kan, Manfei
    Ren, Yonggong
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2025, 16 (02) : 1291 - 1302