Assertion Detection in Clinical Natural Language Processing using Large Language Models

被引:0
|
作者
Ji, Yuelyu [1 ]
Yu, Zeshui [2 ]
Wang, Yanshan [3 ]
机构
[1] Univ Pittsburgh, Dept Comp & Informat, Pittsburgh, PA 15260 USA
[2] Univ Pittsburgh, Dept Pharmaceut Sci, Pittsburgh, PA USA
[3] Univ Pittsburgh, Dept Hlth Informat Management, Pittsburgh, PA USA
基金
美国国家卫生研究院;
关键词
Assertion Detection Large Language Model In-context Learning LoRA Fine-tuning;
D O I
10.1109/ICHI61247.2024.00039
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we aim to address the task of assertion detection when extracting medical concepts from clinical notes, a key process in clinical natural language processing (NLP). Assertion detection in clinical NLP usually involves identifying assertion types for medical concepts in the clinical text, namely certainty (whether the medical concept is positive, negated, possible, or hypothetical), temporality (whether the medical concept is for present or the past history), and experiencer (whether the medical concept is described for the patient or a family member). These assertion types are essential for healthcare professionals to quickly and clearly understand the context of medical conditions from unstructured clinical texts, directly influencing the quality and outcomes of patient care. Although widely used, traditional methods, particularly rule-based NLP systems and machine learning or deep learning models, demand intensive manual efforts to create patterns and tend to overlook less common assertion types, leading to an incomplete understanding of the context. To address this challenge, our research introduces a novel methodology that utilizes Large Language Models (LLMs) pre-trained on a vast array of medical data for assertion detection. We enhanced the current method with advanced reasoning techniques, including Tree of Thought (ToT), Chain of Thought (CoT), and Self-Consistency (SC), and refine it further with Low-Rank Adaptation (LoRA) fine-tuning. We first evaluated the model on the i2b2 2010 assertion dataset. Our method achieved a micro-averaged F-1 of 0.89, with 0.11 improvements over the previous works. To further assess the generalizability of our approach, we extended our evaluation to a local dataset that focused on sleep concept extraction. Our approach achieved an F-1 of 0.74, which is 0.31 higher than the previous method. The results show that using LLMs is a viable option for assertion detection in clinical NLP and can potentially integrate with other LLM-based concept extraction models for clinical NLP tasks.
引用
收藏
页码:242 / 247
页数:6
相关论文
共 50 条
  • [31] Cognitive Impairments in Schizophrenia: A Study in a Large Clinical Sample Using Natural Language Processing
    Mascio, Aurelie
    Stewart, Robert
    Botelle, Riley
    Williams, Marcus
    Mirza, Luwaiza
    Patel, Rashmi
    Pollak, Thomas
    Dobson, Richard
    Roberts, Angus
    FRONTIERS IN DIGITAL HEALTH, 2021, 3
  • [32] Dissecting word embeddings and language models in natural language processing
    Verma, Vivek Kumar
    Pandey, Mrigank
    Jain, Tarun
    Tiwari, Pradeep Kumar
    JOURNAL OF DISCRETE MATHEMATICAL SCIENCES & CRYPTOGRAPHY, 2021, 24 (05): : 1509 - 1515
  • [33] Verbal lie detection using Large Language Models
    Loconte, Riccardo
    Russo, Roberto
    Capuozzo, Pasquale
    Pietrini, Pietro
    Sartori, Giuseppe
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [34] Software Vulnerability Detection using Large Language Models
    Das Purba, Moumita
    Ghosh, Arpita
    Radford, Benjamin J.
    Chu, Bill
    2023 IEEE 34TH INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING WORKSHOPS, ISSREW, 2023, : 112 - 119
  • [35] Explaining Misinformation Detection Using Large Language Models
    Pendyala, Vishnu S.
    Hall, Christopher E.
    ELECTRONICS, 2024, 13 (09)
  • [36] Sarcasm detection in natural language processing
    Ashwitha, A.
    Shruthi, G.
    Shruthi, H. R.
    Upadhyaya, Makarand
    Ray, Abhra Pratip
    Manjunath, T. C.
    MATERIALS TODAY-PROCEEDINGS, 2021, 37 : 3324 - 3331
  • [37] Fraud detection with natural language processing
    Boulieris, Petros
    Pavlopoulos, John
    Xenos, Alexandros
    Vassalos, Vasilis
    MACHINE LEARNING, 2024, 113 (08) : 5087 - 5108
  • [38] Using Natural Language Processing to Identify Stigmatizing Language in Labor and Birth Clinical Notes
    Barcelona, Veronica
    Scharp, Danielle
    Moen, Hans
    Davoudi, Anahita
    Idnay, Betina R.
    Cato, Kenrick
    Topaz, Maxim
    MATERNAL AND CHILD HEALTH JOURNAL, 2023, 28 (3) : 578 - 586
  • [39] Using Natural Language Processing to Identify Stigmatizing Language in Labor and Birth Clinical Notes
    Veronica Barcelona
    Danielle Scharp
    Hans Moen
    Anahita Davoudi
    Betina R. Idnay
    Kenrick Cato
    Maxim Topaz
    Maternal and Child Health Journal, 2024, 28 : 578 - 586
  • [40] Automated anonymization of radiology reports: comparison of publicly available natural language processing and large language models
    Langenbach, Marcel C.
    Foldyna, Borek
    Hadzic, Ibrahim
    Langenbach, Isabel L.
    Raghu, Vineet K.
    Lu, Michael T.
    Neilan, Tomas G.
    Heemelaar, Julius C.
    EUROPEAN RADIOLOGY, 2024,