A few-shot learning method based on knowledge graph in large language models

被引:0
|
作者
Wang, Feilong [1 ,2 ]
Shi, Donghui [1 ,2 ]
Aguilar, Jose [3 ,4 ,5 ]
Cui, Xinyi [1 ]
机构
[1] Anhui Jianzhu Univ, Sch Elect & Informat Engn, Dept Comp Engn, Hefei 230601, Peoples R China
[2] Mass Spectrometry Key Technol R&D&Clin Applicat An, Hefei 230601, Peoples R China
[3] Univ EAFIT, Grp Invest IDI T, Medellin, Colombia
[4] Univ Los Andes, Ctr Estudios Microelect & Sistemas Distribuidos, Merida, Venezuela
[5] IMDEA Networks Inst, Madrid, Spain
关键词
Large language model; Few-shot learning; Fine-tuning; Knowledge-driven dialog generation; Knowledge graph;
D O I
10.1007/s41060-024-00699-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of large language models has significantly transformed natural language processing and text generation. Fine-tuning these models for specific domains enables them to generate answers tailored to the unique requirements of those fields, such as in legal or medical domains. However, these models often perform poorly in few-shot scenarios. Herein, the challenges of data scarcity in fine-tuning large language models in low-sample scenarios were addressed by proposing three different KDGI (Knowledge-Driven Dialog Generation Instances) generation strategies, including entity-based KDGI generation, relation-based KDGI generation, and semantic-based multi-level KDGI generation. These strategies aimed to enhance few-shot datasets to address the issue of low fine-tuning metrics caused by insufficient data. Specifically, knowledge graphs were utilized to define the distinct KDGI generation strategies for enhancing few-shot data. Subsequently, these KDGI data were employed to fine-tune the large language model using the P-tuning v2 approach. Through multiple experiments, the effectiveness of the three KDGI generation strategies was validated using BLEU and ROUGE metrics, and the fine-tuning benefits of few-shot learning on large language models were confirmed. To further evaluate the effectiveness of KDGI, additional experiments were conducted, including LoRA-based fine-tuning in the medical domain and comparative studies leveraging Mask Language Model augmentation, back-translation, and noise injection methods. Consequently, the paper proposes a reference method for leveraging knowledge graphs in prompt data engineering, which shows potential in facilitating few-shot learning for fine-tuning large language models.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Few-shot Knowledge Graph-to-Text Generation with Pretrained Language Models
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Wei, Zhicheng
    Yuan, Nicholas Jing
    Wen, Ji-Rong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1558 - 1568
  • [2] True Few-Shot Learning with Language Models
    Perez, Ethan
    Kiela, Douwe
    Cho, Kyunghyun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [3] Graph Few-Shot Learning via Knowledge Transfer
    Yao, Huaxiu
    Zhang, Chuxu
    Wei, Ying
    Jiang, Meng
    Wang, Suhang
    Huang, Junzhou
    Chawla, Nitesh, V
    Li, Zhenhui
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 6656 - 6663
  • [4] Few-Shot Knowledge Graph Completion Model Based on Relation Learning
    Li, Weijun
    Gu, Jianlai
    Li, Ang
    Gao, Yuxiao
    Zhang, Xinyong
    APPLIED SCIENCES-BASEL, 2023, 13 (17):
  • [5] Large Language Models Enable Few-Shot Clustering
    Viswanathan, Vijay
    Gashteovski, Kiril
    Lawrence, Carolin
    Wu, Tongshuang
    Neubig, Graham
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 321 - 333
  • [6] Few-Shot Knowledge Graph Completion
    Zhang, Chuxu
    Yao, Huaxiu
    Huang, Chao
    Jiang, Meng
    Li, Zhenhui
    Chawla, Nitesh, V
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 3041 - 3048
  • [7] Multimodal Few-Shot Learning with Frozen Language Models
    Tsimpoukelli, Maria
    Menick, Jacob
    Cabi, Serkan
    Eslami, S. M. Ali
    Vinyals, Oriol
    Hill, Felix
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [8] Combat data shift in few-shot learning with knowledge graph
    Zhu, Yongchun
    Zhuang, Fuzhen
    Zhang, Xiangliang
    Qi, Zhiyuan
    Shi, Zhiping
    Cao, Juan
    He, Qing
    FRONTIERS OF COMPUTER SCIENCE, 2023, 17 (01)
  • [9] BayesKGR: Bayesian Few-Shot Learning for Knowledge Graph Reasoning
    Zhao, Feng
    Yan, Cheng
    Jin, Hai
    He, Lifang
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (06)
  • [10] Combat data shift in few-shot learning with knowledge graph
    Yongchun Zhu
    Fuzhen Zhuang
    Xiangliang Zhang
    Zhiyuan Qi
    Zhiping Shi
    Juan Cao
    Qing He
    Frontiers of Computer Science, 2023, 17