Assertion Detection in Clinical Natural Language Processing using Large Language Models

被引:0
|
作者
Ji, Yuelyu [1 ]
Yu, Zeshui [2 ]
Wang, Yanshan [3 ]
机构
[1] Univ Pittsburgh, Dept Comp & Informat, Pittsburgh, PA 15260 USA
[2] Univ Pittsburgh, Dept Pharmaceut Sci, Pittsburgh, PA USA
[3] Univ Pittsburgh, Dept Hlth Informat Management, Pittsburgh, PA USA
基金
美国国家卫生研究院;
关键词
Assertion Detection Large Language Model In-context Learning LoRA Fine-tuning;
D O I
10.1109/ICHI61247.2024.00039
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this study, we aim to address the task of assertion detection when extracting medical concepts from clinical notes, a key process in clinical natural language processing (NLP). Assertion detection in clinical NLP usually involves identifying assertion types for medical concepts in the clinical text, namely certainty (whether the medical concept is positive, negated, possible, or hypothetical), temporality (whether the medical concept is for present or the past history), and experiencer (whether the medical concept is described for the patient or a family member). These assertion types are essential for healthcare professionals to quickly and clearly understand the context of medical conditions from unstructured clinical texts, directly influencing the quality and outcomes of patient care. Although widely used, traditional methods, particularly rule-based NLP systems and machine learning or deep learning models, demand intensive manual efforts to create patterns and tend to overlook less common assertion types, leading to an incomplete understanding of the context. To address this challenge, our research introduces a novel methodology that utilizes Large Language Models (LLMs) pre-trained on a vast array of medical data for assertion detection. We enhanced the current method with advanced reasoning techniques, including Tree of Thought (ToT), Chain of Thought (CoT), and Self-Consistency (SC), and refine it further with Low-Rank Adaptation (LoRA) fine-tuning. We first evaluated the model on the i2b2 2010 assertion dataset. Our method achieved a micro-averaged F-1 of 0.89, with 0.11 improvements over the previous works. To further assess the generalizability of our approach, we extended our evaluation to a local dataset that focused on sleep concept extraction. Our approach achieved an F-1 of 0.74, which is 0.31 higher than the previous method. The results show that using LLMs is a viable option for assertion detection in clinical NLP and can potentially integrate with other LLM-based concept extraction models for clinical NLP tasks.
引用
收藏
页码:242 / 247
页数:6
相关论文
共 50 条
  • [41] Comparison of large language models and traditional natural language processing techniques in predicting arteriovenous fistula failure
    Lama, Suman
    Zhang, Hanjie
    Monaghan, Caitlin
    Bellocchio, Francesco
    Chaudhuri, Sheetal
    Neri, Luca
    Usvyat, Len
    NEPHROLOGY DIALYSIS TRANSPLANTATION, 2024, 39 : I1303 - I1304
  • [42] Accelerated evidence synthesis in orthopaedics—the roles of natural language processing, expert annotation and large language models
    Bálint Zsidai
    Janina Kaarre
    Ann-Sophie Hilkert
    Eric Narup
    Eric Hamrin Senorski
    Alberto Grassi
    Olufemi R. Ayeni
    Volker Musahl
    Christophe Ley
    Elmar Herbst
    Michael T. Hirschmann
    Sebastian Kopf
    Romain Seil
    Thomas Tischer
    Kristian Samuelsson
    Robert Feldt
    Journal of Experimental Orthopaedics, 10
  • [43] Recent Advances in Natural Language Processing via Large Pre-trained Language Models: A Survey
    Min, Bonan
    Ross, Hayley
    Sulem, Elior
    Ben Veyseh, Amir Pouran
    Nguyen, Thien Huu
    Sainz, Oscar
    Agirre, Eneko
    Heintz, Ilana
    Roth, Dan
    ACM COMPUTING SURVEYS, 2024, 56 (02)
  • [44] Evaluation of large language models in natural language processing of PET/CT free-text reports
    Bradshaw, Tyler
    Cho, Steve
    JOURNAL OF NUCLEAR MEDICINE, 2021, 62
  • [45] Utilizing natural language processing and large language models in the diagnosis and prediction of infectious diseases: A systematic review
    Omar, Mahmud
    Brin, Dana
    Glicksberg, Benjamin
    Klang, Eyal
    AMERICAN JOURNAL OF INFECTION CONTROL, 2024, 52 (09) : 992 - 1001
  • [46] Quantifying Uncertainty in Natural Language Explanations of Large Language Models
    Tanneru, Sree Harsha
    Agarwal, Chirag
    Lakkaraju, Himabindu
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [47] Comparison of large language models and traditional natural language processing techniques in predicting arteriovenous fistula failure
    Lama, Suman
    Zhang, Hanjie
    Monaghan, Caitlin
    Bellocchio, Francesco
    Chaudhuri, Sheetal
    Neri, Luca
    Usvyat, Len
    NEPHROLOGY DIALYSIS TRANSPLANTATION, 2024, 39
  • [48] Applications of Natural Language Processing and Large Language Models for Social Determinants of Health: Protocol for a Systematic Review
    Rajwal, Swati
    Zhang, Ziyuan
    Chen, Yankai
    Rogers, Hannah
    Sarker, Abeed
    Xiao, Yunyu
    JMIR RESEARCH PROTOCOLS, 2025, 14
  • [49] Shortcut Learning of Large Language Models in Natural Language Understanding
    Du, Mengnan
    He, Fengxiang
    Zou, Na
    Tao, Dacheng
    Hu, Xia
    COMMUNICATIONS OF THE ACM, 2024, 67 (01) : 110 - 120
  • [50] Processing natural language without natural language processing
    Brill, E
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, PROCEEDINGS, 2003, 2588 : 360 - 369