Exploring the applicability of large language models to citation context analysis

被引:0
|
作者
Nishikawa, Kai [1 ,2 ]
Koshiba, Hitoshi [2 ]
机构
[1] Univ Tsukuba, Inst Lib Informat & Media Sci, 1-2 Kasuga, Tsukuba, Ibaraki 3058550, Japan
[2] Minist Culture Sci & Sports MEXT, Natl Inst Sci & Technol Policy NISTEP, 3-2-2 Kasumigaseki,Chiyoda Ku, Tokyo 1000013, Japan
关键词
Scientometrics; Citation context analysis; Annotation; Large language models (LLM); Generative pre-trained transformer (GPT); COUNTS MEASURE;
D O I
10.1007/s11192-024-05142-9
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Unlike traditional citation analysis, which assumes that all citations in a paper are equivalent, citation context analysis considers the contextual information of individual citations. However, citation context analysis requires creating a large amount of data through annotation, which hinders its widespread use. This study explored the applicability of Large Language Models (LLM)-particularly Generative Pre-trained Transformer (GPT)-to citation context analysis by comparing LLM and human annotation results. The results showed that LLM annotation is as good as or better than human annotation in terms of consistency but poor in terms of its predictive performance. Thus, having LLM immediately replace human annotators in citation context analysis is inappropriate. However, the annotation results obtained by LLM can be used as reference information when narrowing the annotation results obtained by multiple human annotators down to one; alternatively, the LLM can be used as an annotator when it is difficult to prepare sufficient human annotators. This study provides basic findings important for the future development of citation context analysis.
引用
收藏
页码:6751 / 6777
页数:27
相关论文
共 50 条
  • [1] On the Capacity of Citation Generation by Large Language Models
    Qian, Haosheng
    Fan, Yixing
    Zhang, Ruqing
    Guo, Jiafeng
    INFORMATION RETRIEVAL, CCIR 2024, 2025, 15418 : 109 - 123
  • [2] Exploring Distributional Shifts in Large Language Models for Code Analysis
    Arakelyan, Shushan
    Das, Rocktim Jyoti
    Mao, Yi
    Ren, Xiang
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 16298 - 16314
  • [3] Exploring Large Language Models for Classical Philology
    Riemenschneider, Frederick
    Frank, Anette
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15181 - 15199
  • [4] Exploring Mathematical Conjecturing with Large Language Models
    Johansson, Moa
    Smallbone, Nicholas
    NEURAL-SYMBOLIC LEARNING AND REASONING 2023, NESY 2023, 2023,
  • [5] Exploring Length Generalization in Large Language Models
    Anil, Cem
    Wu, Yuhuai
    Andreassen, Anders
    Lewkowycz, Aitor
    Misra, Vedant
    Ramasesh, Vinay
    Slone, Ambrose
    Gur-Ari, Guy
    Dyer, Ethan
    Neyshabur, Behnam
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [6] Applicability of large language models and generative models for legal case judgement summarization
    Deroy, Aniket
    Ghosh, Kripabandhu
    Ghosh, Saptarshi
    ARTIFICIAL INTELLIGENCE AND LAW, 2024,
  • [7] Scientific Software Citation Intent Classification Using Large Language Models
    Istrate, Ana-Maria
    Fisher, Joshua
    Yang, Xinyu
    Moraw, Kara
    Li, Kai
    Li, Donghui
    Klein, Martin
    NATURAL SCIENTIFIC LANGUAGE PROCESSING AND RESEARCH KNOWLEDGE GRAPHS, NSLP 2024, 2024, 14770 : 80 - 99
  • [8] A Concise Review of Long Context in Large Language Models
    Huang, Haitao
    Liang, Zijing
    Fang, Zirui
    Wang, Zhiyuan
    Chen, Mingxiu
    Hong, Yifan
    Liu, Ke
    Shang, Penghui
    PROCEEDINGS OF INTERNATIONAL CONFERENCE ON ALGORITHMS, SOFTWARE ENGINEERING, AND NETWORK SECURITY, ASENS 2024, 2024, : 563 - 566
  • [9] Meta-in-context learning in large language models
    Coda-Forno, Julian
    Binz, Marcel
    Akata, Zeynep
    Botvinick, Matthew
    Wang, Jane X.
    Schulz, Eric
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Context-faithful Prompting for Large Language Models
    Zhou, Wenxuan
    Zhang, Sheng
    Poon, Hoifung
    Chen, Muhao
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 14544 - 14556