Multi-step Transfer Learning in Natural Language Processing for the Health Domain

被引:0
|
作者
Manaka, Thokozile [1 ]
Van Zyl, Terence [2 ]
Kar, Deepak [3 ]
Wade, Alisha [4 ]
机构
[1] Univ Witwatersrand, Sch Comp Sci & Appl Math, Johannesburg, Gauteng, South Africa
[2] Univ Johannesburg, Inst Intelligent Syst, Johannesburg, Gauteng, South Africa
[3] Univ Witwatersrand, Sch Phys, Johannesburg, Gauteng, South Africa
[4] Univ Witwatersrand, Sch Publ Hlth, MRC Wits Rural Publ Hlth & Hlth Transit Res Unit, Johannesburg, South Africa
关键词
Transfer learning; Verbal autopsy; Natural language processing; Text classification; Feature extraction; Fine tuning; MODEL;
D O I
10.1007/s11063-024-11526-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The restricted access to data in healthcare facilities due to patient privacy and confidentiality policies has led to the application of general natural language processing (NLP) techniques advancing relatively slowly in the health domain. Additionally, because clinical data is unique to various institutions and laboratories, there are not enough standards and conventions for data annotation. In places without robust death registration systems, the cause of death (COD) is determined through a verbal autopsy (VA) report. A non-clinician field agent completes a VA report using a set of standardized questions as guide to identify the symptoms of a COD. The narrative text of the VA report is used as a case study to examine the difficulties of applying NLP techniques to the healthcare domain. This paper presents a framework that leverages knowledge across multiple domains via two domain adaptation techniques: feature extraction and fine-tuning. These techniques aim to improve VA text representations for COD classification tasks in the health domain. The framework is motivated by multi-step learning, where a final learning task is realized via a sequence of intermediate learning tasks. The framework builds upon the strengths of the Bidirectional Encoder Representations from Transformers (BERT) and Embeddings from Language Models (ELMo) models pretrained on the general English and biomedical domains. These models are employed to extract features from the VA narratives. Our results demonstrate improved performance when initializing the learning of BERT embeddings with ELMo embeddings. The benefit of incorporating character-level information for learning word embeddings in the English domain, coupled with word-level information for learning word embeddings in the biomedical domain, is also evident.
引用
收藏
页数:26
相关论文
共 50 条
  • [11] Genetic Multi-Step Search in Interpolation and Extrapolation Domain
    Hanada, Yoshiko
    Hiroyasu, Tomoyuki
    Mitsunori, Miki
    GECCO 2007: GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, VOL 1 AND 2, 2007, : 1242 - 1249
  • [12] Incremental multi-step Q-learning
    Peng, J
    Williams, RJ
    MACHINE LEARNING, 1996, 22 (1-3) : 283 - 290
  • [13] Multi-Task Learning in Natural Language Processing: An Overview
    Chen, Shijie
    Zhang, Yu
    Yang, Qiang
    ACM COMPUTING SURVEYS, 2024, 56 (12)
  • [14] Multi-Step Reinforcement Learning: A Unifying Algorithm
    De Asis, Kristopher
    Hernandez-Garcia, J. Fernando
    Holland, G. Zacharias
    Sutton, Richard S.
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 2902 - 2909
  • [15] Learning-assisted multi-step planning
    Hauser, K
    Bretl, T
    Latombe, JC
    2005 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-4, 2005, : 4575 - 4580
  • [16] Reinforcement Learning for Multi-Step Expert Advice
    Philipp, Patrick
    Rettinger, Achim
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 962 - 971
  • [17] DoubleTransfer at MEDIQA 2019: Multi-Source Transfer Learning for Natural Language Understanding in the Medical Domain
    Xu, Yichong
    Liu, Xiaodong
    Li, Chunyuan
    Poon, Hoifung
    Gao, Jianfeng
    SIGBIOMED WORKSHOP ON BIOMEDICAL NATURAL LANGUAGE PROCESSING (BIONLP 2019), 2019, : 399 - 405
  • [18] Semi-Supervised Learning and Domain Adaptation in Natural Language Processing
    Foster, George
    COMPUTATIONAL LINGUISTICS, 2014, 40 (02) : 519 - 522
  • [19] Semi-Supervised Learning and Domain Adaptation in Natural Language Processing
    Lefevre, Fabrice
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2013, 54 (01): : 235 - 237