Frozen Language Model Helps ECG Zero-Shot Learning

被引:0
|
作者
Li, Jun [1 ]
Liu, Che [2 ,3 ]
Cheng, Sibo [3 ]
Arcucci, Rossella [2 ,3 ]
Hong, Shenda [4 ,5 ]
机构
[1] Jilin Univ, Coll Elect Sci & Engn, Changchun, Peoples R China
[2] Imperial Coll London, Dept Earth Sci & Engn, London SW7 2AZ, England
[3] Imperial Coll London, Data Sci Inst, Dept Comp, London, England
[4] Peking Univ, Natl Inst Hlth Data Sci, Beijing, Peoples R China
[5] Peking Univ, Inst Med Technol, Hlth Sci Ctr, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimodal self-supervised learning; Zero-shot learning; Language model; ECG; Signal processing; MYOCARDIAL-INFARCTION; SIGNALS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The electrocardiogram (ECG) is one of the most commonly used non-invasive, convenient medical monitoring tools that assist in the clinical diagnosis of heart diseases. Recently, deep learning (DL) techniques, particularly self-supervised learning (SSL), have demonstrated great potential in the classification of ECG. SSL pre-training has achieved competitive performance with only a small amount of annotated data after fine-tuning. However, current SSL methods rely on the availability of annotated data and are unable to predict labels not existing in fine-tuning datasets. To address this challenge, we propose Multimodal ECG-Text Self-supervised pre-training (METS), the first work to utilize the auto-generated clinical reports to guide ECG SSL pre-training. We use a trainable ECG encoder and a frozen language model to embed paired ECG and automatically machine-generated clinical reports separately. The SSL aims to maximize the similarity between paired ECG and auto-generated report while minimize the similarity between ECG and other reports. In downstream classification tasks, METS achieves around 10% improvement in performance without using any annotated data via zero-shot classification, compared to other supervised and SSL baselines that rely on annotated data. Furthermore, METS achieves the highest recall and F1 scores on the MIT-BIH dataset, despite MIT-BIH containing different classes of ECG compared to the pre-trained dataset. The extensive experiments have demonstrated the advantages of using ECG-Text multimodal self-supervised learning in terms of generalizability, effectiveness, and efficiency.
引用
收藏
页码:402 / 415
页数:14
相关论文
共 50 条
  • [1] A Joint Generative Model for Zero-Shot Learning
    Gao, Rui
    Hou, Xingsong
    Qin, Jie
    Liu, Li
    Zhu, Fan
    Zhang, Zhao
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT IV, 2019, 11132 : 631 - 646
  • [2] Model Selection for Generalized Zero-Shot Learning
    Zhang, Hongguang
    Koniusz, Piotr
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT II, 2019, 11130 : 198 - 204
  • [3] Learning a Deep Embedding Model for Zero-Shot Learning
    Zhang, Li
    Xiang, Tao
    Gong, Shaogang
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3010 - 3019
  • [4] Zero-Shot Video Question Answering via Frozen Bidirectional Language Models
    Yang, Antoine
    Miech, Antoine
    Sivic, Josef
    Laptev, Ivan
    Schmid, Cordelia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [5] Zero-shot Learning for Grapheme to Phoneme Conversion with Language Ensemble
    Li, Xinjian
    Metze, Florian
    Mortensen, David R.
    Watanabe, Shinji
    Black, Alan W.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 2106 - 2115
  • [6] Zero-shot Learning of Classifiers from Natural Language Quantification
    Srivastava, Shashank
    Labutov, Igor
    Mitchell, Tom
    PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, 2018, : 306 - 316
  • [7] Ordinal Zero-Shot Learning
    Huo, Zengwei
    Geng, Xin
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1916 - 1922
  • [8] Zero-Shot Kernel Learning
    Zhang, Hongguang
    Koniusz, Piotr
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7670 - 7679
  • [9] Zero-shot causal learning
    Nilforoshan, Hamed
    Moor, Michael
    Roohani, Yusuf
    Chen, Yining
    Surina, Anja
    Yasunaga, Michihiro
    Oblak, Sara
    Leskovec, Jure
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Can a zero-shot learning Large Language Model code complex interview data?
    Balt, E.
    Salmi, S.
    Bhulai, S.
    Eikelenboom, M.
    Gilissen, R.
    Creemers, D.
    Popma, A.
    Merelle, S.
    EUROPEAN JOURNAL OF PUBLIC HEALTH, 2024, 34