The Use of Clinical Language Models Pretrained on Institutional EHR Data for Downstream Tasks

被引:0
|
作者
Suvirat, Kerdkiat [1 ]
Chairat, Sawrawit [1 ]
Horsiritham, Kanakorn [2 ]
Ingviya, Thammasin [3 ]
Kongkamol, Chanon [3 ]
Chaichulee, Sitthichok [1 ]
机构
[1] Prince Songkla Univ, Dept Biomed Sci & Biomed Engn, Fac Med, Hat Yai, Thailand
[2] Prince Songkla Univ, Coll Digital Sci, Hat Yai, Thailand
[3] Prince Songkla Univ, Fac Med, Dept Family & Prevent Med, Hat Yai, Thailand
关键词
natural language processing; language modelling; clinical note; electronic health records; text classification;
D O I
10.1109/JCSSE61278.2024.10613630
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Clinical language models have attracted considerable attention in recent years because of their potential to improve healthcare workflows a nd o ptimise p atient c are. While many pre-trained clinical language models have been published, there are no models specifically f or t he T hai c linical context, where English terminologies are used for diseases, procedures and medications, and Thai language is used for clinical notes. This study investigated the pretraining of different language model architectures, namely RoBERTa, GPT-2 and T5, on the EHR data of Songklanagarind Hospital in Thailand, which includes over 80 million documents. We also investigated the applications of the pretrained model to three downstream clinical tasks: tuberculosis case finding, B IRADS c ategory c lassification an d intraocular pressure extraction. The results indicate that our domain-specific language models performed better than the general-purpose language model, mBERT, and required fewer training examples to achieve the same performance. The study encourages the use of clinical language models to streamline clinical workflows, support clinical research and assist hospital auditing.
引用
收藏
页码:648 / 655
页数:8
相关论文
共 50 条
  • [41] Clinical information extraction for lower-resource languages and domains with few-shot learning using pretrained language models and prompting
    Richter-Pechanski, Phillip
    Wiesenbach, Philipp
    Schwab, Dominic Mathias
    Kiriakou, Christina
    Geis, Nicolas
    Dieterich, Christoph
    Frank, Anette
    NATURAL LANGUAGE PROCESSING, 2024,
  • [42] Optimizing data integration in trials that use EHR data: lessons learned from a multi-center randomized clinical trial
    Raman, Sudha R.
    Qualls, Laura G.
    Hammill, Bradley G.
    Nelson, Adam J.
    Nilles, Ester Kim
    Marsolo, Keith
    O'Brien, Emily C.
    TRIALS, 2023, 24 (01)
  • [43] Exploring the Performance of Large Language Models for Data Analysis Tasks Through the CRISP-DM Framework
    Musazade, Nurlan
    Mezei, Jozsef
    Wang, Xiaolu
    GOOD PRACTICES AND NEW PERSPECTIVES IN INFORMATION SYSTEMS AND TECHNOLOGIES, VOL 5, WORLDCIST 2024, 2024, 989 : 56 - 65
  • [44] Lingdan: enhancing encoding of traditional Chinese medicine knowledge for clinical reasoning tasks with large language models
    Hua, Rui
    Dong, Xin
    Wei, Yu
    Shu, Zixin
    Yang, Pengcheng
    Hu, Yunhui
    Zhou, Shuiping
    Sun, He
    Yan, Kaijing
    Yan, Xijun
    Chang, Kai
    Li, Xiaodong
    Bai, Yuning
    Zhang, Runshun
    Wang, Wenjia
    Zhou, Xuezhong
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024,
  • [45] Am I a Resource-Poor Language? Data Sets, Embeddings, Models and Analysis for four different NLP Tasks in Telugu Language
    Marreddy, Mounika
    Oota, Subba Reddy
    Vakada, Lakshmi Sireesha
    Chinni, Venkata Charan
    Mamidi, Radhika
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (01)
  • [46] ESSENTIAL HYPERTENSION - APPROACH TO CLINICAL-DATA BY THE USE OF MODELS
    CHAU, NP
    SAFAR, ME
    LONDON, GM
    WEISS, YA
    HYPERTENSION, 1979, 1 (02) : 86 - 97
  • [47] Enhancing Clinical Relevance of Pretrained Language Models Through Integration of External Knowledge: Case Study on Cardiovascular Diagnosis From Electronic Health Records
    Lu, Qiuhao
    Wen, Andrew
    Nguyen, Thien
    Liu, Hongfang
    JMIR AI, 2024, 3
  • [48] Recurrent Deep Network Models for Clinical NLP Tasks: Use Case with Sentence Boundary Disambiguation
    Knoll, Benjamin C.
    Lindemann, Elizabeth A.
    Albert, Arian L.
    Melton, Genevieve B.
    Pakhomov, Serguei V. S.
    MEDINFO 2019: HEALTH AND WELLBEING E-NETWORKS FOR ALL, 2019, 264 : 198 - 202
  • [49] Equipping Language Models with Tool Use Capability for Tabular Data Analysis in Finance
    Theuma, Adrian
    Shareghi, Ehsan
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 90 - 103
  • [50] Evaluating the Effectiveness of Large Language Models in Converting Clinical Data to FHIR Format
    Delaunay, Julien
    Girbes, Daniel
    Cusido, Jordi
    APPLIED SCIENCES-BASEL, 2025, 15 (06):