Zero-shot interpretable phenotyping of postpartum hemorrhage using large language models

被引:15
|
作者
Alsentzer, Emily [1 ]
Rasmussen, Matthew J. [2 ]
Fontoura, Romy [2 ]
Cull, Alexis L. [2 ]
Beaulieu-Jones, Brett [3 ]
Gray, Kathryn J. [4 ,5 ]
Bates, David W. [1 ,6 ]
Kovacheva, Vesela P. [2 ]
机构
[1] Brigham & Womens Hosp, Div Gen Internal Med & Primary Care, Boston, MA USA
[2] Brigham & Womens Hosp, Dept Anesthesiol Perioperat & Pain Med, Boston, MA 02115 USA
[3] Univ Chicago, Dept Med, Sect Biomed Data Sci, Chicago, IL USA
[4] Massachusetts Gen Hosp, Ctr Genom Med, Boston, MA USA
[5] Brigham & Womens Hosp, Div Maternal Fetal Med, Boston, MA USA
[6] Harvard TH Chan Sch Publ Hlth, Dept Hlth Care Policy & Management, Boston, MA USA
关键词
CLASSIFICATION; ALGORITHMS;
D O I
10.1038/s41746-023-00957-x
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Many areas of medicine would benefit from deeper, more accurate phenotyping, but there are limited approaches for phenotyping using clinical notes without substantial annotated data. Large language models (LLMs) have demonstrated immense potential to adapt to novel tasks with no additional training by specifying task-specific instructions. Here we report the performance of a publicly available LLM, Flan-T5, in phenotyping patients with postpartum hemorrhage (PPH) using discharge notes from electronic health records (n = 271,081). The language model achieves strong performance in extracting 24 granular concepts associated with PPH. Identifying these granular concepts accurately allows the development of interpretable, complex phenotypes and subtypes. The Flan-T5 model achieves high fidelity in phenotyping PPH (positive predictive value of 0.95), identifying 47% more patients with this complication compared to the current standard of using claims codes. This LLM pipeline can be used reliably for subtyping PPH and outperforms a claims-based approach on the three most common PPH subtypes associated with uterine atony, abnormal placentation, and obstetric trauma. The advantage of this approach to subtyping is its interpretability, as each concept contributing to the subtype determination can be evaluated. Moreover, as definitions may change over time due to new guidelines, using granular concepts to create complex phenotypes enables prompt and efficient updating of the algorithm. Using this language modelling approach enables rapid phenotyping without the need for any manually annotated training data across multiple clinical use cases.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Large Language Models Are Zero-Shot Fuzzers: Fuzzing Deep-Learning Libraries via Large Language Models
    Deng, Yinlin
    Xia, Chunqiu Steven
    Peng, Haoran
    Yang, Chenyuan
    Zhan, Lingming
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 423 - 435
  • [22] Enhancing Zero-shot Audio Classification using Sound Attribute Knowledge from Large Language Models
    Xu, Xuenan
    Zhang, Pingyue
    Yang, Ming
    Zhang, Ji
    Wu, Mengyue
    INTERSPEECH 2024, 2024, : 4808 - 4812
  • [23] Harnessing large language models' zero-shot and few-shot learning capabilities for regulatory research
    Meshkin, Hamed
    Zirkle, Joel
    Arabidarrehdor, Ghazal
    Chaturbedi, Anik
    Chakravartula, Shilpa
    Mann, John
    Thrasher, Bradlee
    Li, Zhihua
    BRIEFINGS IN BIOINFORMATICS, 2024, 25 (05)
  • [25] Zero-Shot ECG Diagnosis with Large Language Models and Retrieval-Augmented Generation
    Yu, Han
    Guo, Peikun
    Sano, Akane
    MACHINE LEARNING FOR HEALTH, ML4H, VOL 225, 2023, 225 : 650 - 663
  • [26] Aligning Instruction Tasks Unlocks Large Language Models as Zero-Shot Relation Extractors
    Zhang, Kai
    Gutierrez, Bernal Jimenez
    Su, Yu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 794 - 812
  • [27] ZVQAF: Zero-shot visual question answering with feedback from large language models
    Liu, Cheng
    Wang, Chao
    Peng, Yan
    Li, Zhixu
    NEUROCOMPUTING, 2024, 580
  • [28] The unreasonable effectiveness of large language models in zero-shot semantic annotation of legal texts
    Savelka, Jaromir
    Ashley, Kevin D.
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2023, 6
  • [30] Improving Zero-Shot Stance Detection by Infusing Knowledge from Large Language Models
    Guo, Mengzhuo
    Jiang, Xiaorui
    Liao, Yong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT XIII, ICIC 2024, 2024, 14874 : 121 - 132