Multi-step Transfer Learning in Natural Language Processing for the Health Domain

被引:0
|
作者
Manaka, Thokozile [1 ]
Van Zyl, Terence [2 ]
Kar, Deepak [3 ]
Wade, Alisha [4 ]
机构
[1] Univ Witwatersrand, Sch Comp Sci & Appl Math, Johannesburg, Gauteng, South Africa
[2] Univ Johannesburg, Inst Intelligent Syst, Johannesburg, Gauteng, South Africa
[3] Univ Witwatersrand, Sch Phys, Johannesburg, Gauteng, South Africa
[4] Univ Witwatersrand, Sch Publ Hlth, MRC Wits Rural Publ Hlth & Hlth Transit Res Unit, Johannesburg, South Africa
关键词
Transfer learning; Verbal autopsy; Natural language processing; Text classification; Feature extraction; Fine tuning; MODEL;
D O I
10.1007/s11063-024-11526-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The restricted access to data in healthcare facilities due to patient privacy and confidentiality policies has led to the application of general natural language processing (NLP) techniques advancing relatively slowly in the health domain. Additionally, because clinical data is unique to various institutions and laboratories, there are not enough standards and conventions for data annotation. In places without robust death registration systems, the cause of death (COD) is determined through a verbal autopsy (VA) report. A non-clinician field agent completes a VA report using a set of standardized questions as guide to identify the symptoms of a COD. The narrative text of the VA report is used as a case study to examine the difficulties of applying NLP techniques to the healthcare domain. This paper presents a framework that leverages knowledge across multiple domains via two domain adaptation techniques: feature extraction and fine-tuning. These techniques aim to improve VA text representations for COD classification tasks in the health domain. The framework is motivated by multi-step learning, where a final learning task is realized via a sequence of intermediate learning tasks. The framework builds upon the strengths of the Bidirectional Encoder Representations from Transformers (BERT) and Embeddings from Language Models (ELMo) models pretrained on the general English and biomedical domains. These models are employed to extract features from the VA narratives. Our results demonstrate improved performance when initializing the learning of BERT embeddings with ELMo embeddings. The benefit of incorporating character-level information for learning word embeddings in the English domain, coupled with word-level information for learning word embeddings in the biomedical domain, is also evident.
引用
收藏
页数:26
相关论文
共 50 条
  • [21] Protocols for multi-step thermoplastic processing of metallic glasses
    Bordeenithikasem, Punnathat
    Sohn, Sungwoo
    Liu, Ze
    Schroers, Jan
    SCRIPTA MATERIALIA, 2015, 104 : 56 - 59
  • [22] GTLNLP: A Mathematical Exploration of Cross-Domain Knowledge Transfer for Text Generation for Generative Transfer Learning in Natural Language Processing
    Patil, Swati Bula
    Talekar, Sopan
    Vyawahare, Mohini
    Bhosle, Amol A.
    Bramhe, Manoj Vasantrao
    Kanwade, Archana Bajirao
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (01) : 379 - 385
  • [23] Flow chemistry in the multi-step synthesis of natural products
    Wan, Li
    Kong, Gaopan
    Liu, Minjie
    Jiang, Meifen
    Cheng, Dang
    Chen, Fener
    GREEN SYNTHESIS AND CATALYSIS, 2022, 3 (03): : 243 - 258
  • [24] Multi-step Forecasting via Multi-task Learning
    Jawed, Shayan
    Rashed, Ahmed
    Schmidt-Thieme, Lars
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 790 - 799
  • [25] Multi-step Prediction for Learning Invariant Representations in Reinforcement Learning
    Xu, Xinyue
    Lv, Kai
    Dong, Xingye
    Han, Sheng
    Lin, Youfang
    2021 INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE BIG DATA AND INTELLIGENT SYSTEMS (HPBD&IS), 2021, : 202 - 206
  • [26] An Evaluation of Progressive Neural Networks for Transfer Learning in Natural Language Processing
    Hagerer, Gerhard
    Moeed, Abdul
    Dugar, Sumit
    Gupta, Sarthak
    Ghosh, Mainak
    Danner, Hannah
    Mitevski, Oliver
    Nawroth, Andreas
    Groh, Georg
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 1376 - 1381
  • [27] Multi-Step Deductive Reasoning Over Natural Language: An Empirical Study on Out-of-Distribution Generalisation
    Bao, Qiming
    Peng, Alex Yuxuan
    Hartill, Tim
    Tan, Neset
    Deng, Zhenyun
    Witbrock, Michael
    Liu, Jiamou
    NEURAL-SYMBOLIC LEARNING AND REASONING, NESY 2022, 2022, : 202 - 217
  • [28] Distilling Multi-Step Reasoning Capabilities into Smaller Language Model
    Yim, Yauwai
    Wang, Zirui
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 530 - 535
  • [29] Effectiveness of Multi-Step Crossover in Extrapolation Domain for Genetic Programming
    Kuroda, Mao
    Hanada, Yoshiko
    Ono, Keiko
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 3654 - 3659
  • [30] Analysis of multi-step algorithms for cognitive maps learning
    Jastriebow, A.
    Poczeta, K.
    BULLETIN OF THE POLISH ACADEMY OF SCIENCES-TECHNICAL SCIENCES, 2014, 62 (04) : 735 - 741