SikuGPT: A Generative Pre-trained Model for Intelligent Information Processing of Ancient Texts from the Perspective of Digital Humanities

被引:1
|
作者
Liu, Chang [1 ]
Wang, Dongbo [1 ]
Zhao, Zhixiao [1 ]
Hu, Die [1 ]
Wu, Mengcheng [1 ]
Zhang, Hai [1 ]
Lin, Li tao [2 ]
Liu, Jiangfeng [2 ]
Shen, Si [3 ]
Li, Bin [4 ]
Zhao, Lianzhen [5 ]
机构
[1] Nanjing Agr Univ, Coll Informat Management, Nanjing, Peoples R China
[2] Nanjing Univ, Sch Informat Management, Nanjing, Peoples R China
[3] Nanjing Univ Sci & Technol, Sch Econ & Management, Grp Sci & Technol Full Text Knowledge Min, Nanjing, Peoples R China
[4] Nanjing Normal Univ, Coll Liberal Art, Nanjing, Peoples R China
[5] China Pharmaceut Univ, Sch Foreign Languages, Nanjing, Peoples R China
来源
关键词
Generative pre-trained model; Siku Quanshu; Chinese ancient texts; digital humanities research; natural language processing;
D O I
10.1145/3676969
中图分类号
C [社会科学总论];
学科分类号
03 ; 0303 ;
摘要
The rapid development of generative artificial intelligence has brought significant opportunities for the advancement of digital humanities research. Intelligent processing of ancient texts, as an essential part of digital humanities, is also undergoing a transformation in research methodologies in the wave of AIGC. The integration of generative pre-trained models with Chinese ancient texts, a vital carrier of Chinese culture, allows for deep mining of the content of these texts and provides services that make ancient texts more understandable and accessible to the general public. In this research, we propose a method that combines the most renowned Chinese anthology, the "Siku Quanshu," with generative pre-trained models. We developed the SikuGPT model, a generative model for ancient text processing tasks, based on GPT-type language models by continued pretraining. This model was tested on two typical tasks of ancient text processing: translation between classical and modern Chinese, and classification of ancient texts. The findings reveal that our model achieves advantages in understanding and generating scenarios of ancient texts. The capability of SikuGPT in processing traditional Chinese texts helps to promote the organization of ancient information and knowledge services, and advances the international dissemination of traditional Chinese culture.
引用
收藏
页数:17
相关论文
共 11 条
  • [1] Intelligent Completion of Ancient Texts Based on Pre-trained Language Models
    Li J.
    Ming C.
    Guo Z.
    Qian T.
    Peng Z.
    Wang X.
    Li X.
    Li J.
    Data Analysis and Knowledge Discovery, 2024, 8 (05) : 59 - 67
  • [2] A text restoration model for ancient texts based on pre-trained language model RoBERTa
    Gu, Zhongyu
    Guan, Yanzhi
    Zhang, Shuai
    PROCEEDINGS OF 2024 4TH INTERNATIONAL CONFERENCE ON INTERNET OF THINGS AND MACHINE LEARNING, IOTML 2024, 2024, : 96 - 102
  • [3] A reinforcement learning intelligent deductive model with pre-trained sequence information
    Han, Xinyu
    Xu, Huosheng
    Yu, Hao
    Li, Sizhao
    INTERNATIONAL JOURNAL OF BIO-INSPIRED COMPUTATION, 2023, 22 (04) : 195 - 205
  • [4] Fusing graph structural information with pre-trained generative model for knowledge graph-to-text generation
    Shi, Xiayang
    Xia, Zhenlin
    Li, Yinlin
    Wang, Xuhui
    Niu, Yufeng
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (03) : 2619 - 2640
  • [5] iProL: identifying DNA promoters from sequence information based on Longformer pre-trained model
    Peng, Binchao
    Sun, Guicong
    Fan, Yongxian
    BMC BIOINFORMATICS, 2024, 25 (01):
  • [6] Extracting structured information from unstructured histopathology reports using generative pre-trained transformer 4 (GPT-4)
    Truhn, Daniel
    Loeffler, Chiara M. L.
    Mueller-Franzes, Gustav
    Nebelung, Sven
    Hewitt, Katherine J.
    Brandner, Sebastian
    Bressem, Keno K.
    Foersch, Sebastian
    Kather, Jakob Nikolas
    JOURNAL OF PATHOLOGY, 2024, 262 (03): : 310 - 319
  • [7] BROS: A Pre-trained Language Model Focusing on Text and Layout for Better Key Information Extraction from Documents
    Hong, Teakgyu
    Kim, Donghyun
    Ji, Mingi
    Hwang, Wonseok
    Nam, Daehyun
    Park, Sungrae
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10767 - 10775
  • [8] Improving Pre-Trained Model-Based Speech Emotion Recognition From a Low-Level Speech Feature Perspective
    Liu, Ke
    Wei, Jiwei
    Zou, Jie
    Wang, Peng
    Yang, Yang
    Shen, Heng Tao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 10623 - 10636
  • [9] Development and evaluation of a program based on a generative pre-trained transformer model from a public natural language processing platform for efficiency enhancement in post-procedural quality control of esophageal endoscopic submucosal dissection
    Ma, Huaiyuan
    Ma, Xingbin
    Yang, Chunxiao
    Niu, Qiong
    Gao, Tao
    Liu, Chengxia
    Chen, Yan
    SURGICAL ENDOSCOPY AND OTHER INTERVENTIONAL TECHNIQUES, 2024, 38 (03): : 1264 - 1272
  • [10] Development and evaluation of a program based on a generative pre-trained transformer model from a public natural language processing platform for efficiency enhancement in post-procedural quality control of esophageal endoscopic submucosal dissection
    Huaiyuan Ma
    Xingbin Ma
    Chunxiao Yang
    Qiong Niu
    Tao Gao
    Chengxia Liu
    Yan Chen
    Surgical Endoscopy, 2024, 38 : 1264 - 1272