Multi-step Transfer Learning in Natural Language Processing for the Health Domain

被引:0
|
作者
Manaka, Thokozile [1 ]
Van Zyl, Terence [2 ]
Kar, Deepak [3 ]
Wade, Alisha [4 ]
机构
[1] Univ Witwatersrand, Sch Comp Sci & Appl Math, Johannesburg, Gauteng, South Africa
[2] Univ Johannesburg, Inst Intelligent Syst, Johannesburg, Gauteng, South Africa
[3] Univ Witwatersrand, Sch Phys, Johannesburg, Gauteng, South Africa
[4] Univ Witwatersrand, Sch Publ Hlth, MRC Wits Rural Publ Hlth & Hlth Transit Res Unit, Johannesburg, South Africa
关键词
Transfer learning; Verbal autopsy; Natural language processing; Text classification; Feature extraction; Fine tuning; MODEL;
D O I
10.1007/s11063-024-11526-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The restricted access to data in healthcare facilities due to patient privacy and confidentiality policies has led to the application of general natural language processing (NLP) techniques advancing relatively slowly in the health domain. Additionally, because clinical data is unique to various institutions and laboratories, there are not enough standards and conventions for data annotation. In places without robust death registration systems, the cause of death (COD) is determined through a verbal autopsy (VA) report. A non-clinician field agent completes a VA report using a set of standardized questions as guide to identify the symptoms of a COD. The narrative text of the VA report is used as a case study to examine the difficulties of applying NLP techniques to the healthcare domain. This paper presents a framework that leverages knowledge across multiple domains via two domain adaptation techniques: feature extraction and fine-tuning. These techniques aim to improve VA text representations for COD classification tasks in the health domain. The framework is motivated by multi-step learning, where a final learning task is realized via a sequence of intermediate learning tasks. The framework builds upon the strengths of the Bidirectional Encoder Representations from Transformers (BERT) and Embeddings from Language Models (ELMo) models pretrained on the general English and biomedical domains. These models are employed to extract features from the VA narratives. Our results demonstrate improved performance when initializing the learning of BERT embeddings with ELMo embeddings. The benefit of incorporating character-level information for learning word embeddings in the English domain, coupled with word-level information for learning word embeddings in the biomedical domain, is also evident.
引用
收藏
页数:26
相关论文
共 50 条
  • [1] Multi-Step Transfer Learning for Sentiment Analysis
    Golubev, Anton
    Loukachevitch, Natalia
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2021), 2021, 12801 : 209 - 217
  • [2] Transfer Learning for Multi-Step Resource Utilization Prediction
    Parera, Claudia
    Liao, Qi
    Malanchini, Ilaria
    Wellington, Dan
    Redondi, Alessandro E. C.
    Cesana, Matteo
    2020 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (IEEE PIMRC), 2020,
  • [3] Multi-step Iterative Automated Domain Modeling with Large Language Models
    Yang, Yujing
    Chen, Boqi
    Chen, Kua
    Mussbacher, Gunter
    Varro, Daniel
    ACM/IEEE 27TH INTERNATIONAL CONFERENCE ON MODEL DRIVEN ENGINEERING LANGUAGES AND SYSTEMS: COMPANION PROCEEDINGS, MODELS 2024, 2024, : 587 - 595
  • [4] MULTI-STEP SPOKEN LANGUAGE UNDERSTANDING SYSTEM BASED ON ADVERSARIAL LEARNING
    Wang, Yu
    Shen, Yilin
    Jin, Hongxia
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7673 - 7677
  • [5] Fully Convolutional Network with Multi-Step Reinforcement Learning for Image Processing
    Furuta, Ryosuke
    Inoue, Naoto
    Yamasaki, Toshihiko
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 3598 - 3605
  • [6] MULTI-STEP ONLINE UNSUPERVISED DOMAIN ADAPTATION
    Moon, J. H.
    Das, Debasnnt
    Lee, C. S. George
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 4172 - 4176
  • [7] A Multi-Step Reinforcement Learning Algorithm
    Zhang, Zhicong
    Hu, Kaishun
    Huang, Huiyu
    Li, Shuai
    Zhao, Shaoyong
    FRONTIERS OF MANUFACTURING AND DESIGN SCIENCE, PTS 1-4, 2011, 44-47 : 3611 - 3615
  • [8] Processing of IDS alerts in multi-step attacks
    Bajtos, Tomas
    Sokol, Pavol
    Kurimsky, Frantisek
    SOFTWARE IMPACTS, 2024, 19
  • [9] Multi-domain spoken language understanding with transfer learning
    Jeong, Minwoo
    Lee, Gary Geunbae
    SPEECH COMMUNICATION, 2009, 51 (05) : 412 - 424
  • [10] Multi-step truncated Q learning algorithm
    Chen, SL
    Wu, HZ
    Han, XL
    Xiao, L
    PROCEEDINGS OF 2005 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-9, 2005, : 194 - 198