Multi-step Transfer Learning in Natural Language Processing for the Health Domain

被引:0
|
作者
Manaka, Thokozile [1 ]
Van Zyl, Terence [2 ]
Kar, Deepak [3 ]
Wade, Alisha [4 ]
机构
[1] Univ Witwatersrand, Sch Comp Sci & Appl Math, Johannesburg, Gauteng, South Africa
[2] Univ Johannesburg, Inst Intelligent Syst, Johannesburg, Gauteng, South Africa
[3] Univ Witwatersrand, Sch Phys, Johannesburg, Gauteng, South Africa
[4] Univ Witwatersrand, Sch Publ Hlth, MRC Wits Rural Publ Hlth & Hlth Transit Res Unit, Johannesburg, South Africa
关键词
Transfer learning; Verbal autopsy; Natural language processing; Text classification; Feature extraction; Fine tuning; MODEL;
D O I
10.1007/s11063-024-11526-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The restricted access to data in healthcare facilities due to patient privacy and confidentiality policies has led to the application of general natural language processing (NLP) techniques advancing relatively slowly in the health domain. Additionally, because clinical data is unique to various institutions and laboratories, there are not enough standards and conventions for data annotation. In places without robust death registration systems, the cause of death (COD) is determined through a verbal autopsy (VA) report. A non-clinician field agent completes a VA report using a set of standardized questions as guide to identify the symptoms of a COD. The narrative text of the VA report is used as a case study to examine the difficulties of applying NLP techniques to the healthcare domain. This paper presents a framework that leverages knowledge across multiple domains via two domain adaptation techniques: feature extraction and fine-tuning. These techniques aim to improve VA text representations for COD classification tasks in the health domain. The framework is motivated by multi-step learning, where a final learning task is realized via a sequence of intermediate learning tasks. The framework builds upon the strengths of the Bidirectional Encoder Representations from Transformers (BERT) and Embeddings from Language Models (ELMo) models pretrained on the general English and biomedical domains. These models are employed to extract features from the VA narratives. Our results demonstrate improved performance when initializing the learning of BERT embeddings with ELMo embeddings. The benefit of incorporating character-level information for learning word embeddings in the English domain, coupled with word-level information for learning word embeddings in the biomedical domain, is also evident.
引用
收藏
页数:26
相关论文
共 50 条
  • [41] Multi-Agent Reinforcement Learning with Multi-Step Generative Models
    Krupnik, Orr
    Mordatch, Igor
    Tamar, Aviv
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [42] A novel approach to ultra-short-term multi-step wind power predictions based on encoder-decoder architecture in natural language processing
    Wang, Lei
    He, Yigang
    Li, Lie
    Liu, Xiaoyan
    Zhao, Yingying
    JOURNAL OF CLEANER PRODUCTION, 2022, 354
  • [43] Multi-Step and Switchable Energy Transfer in Photoluminescent Organosilicone Capsules
    Yu, Longyue
    Liu, Hailong
    Feng, Ning
    Yi, Gang
    Xin, Xia
    Hao, Jingcheng
    Li, Hongguang
    ADVANCED SCIENCE, 2024, 11 (31)
  • [44] Natural language processing in the patent domain: a survey
    Lekang Jiang
    Stephan M. Goetz
    Artificial Intelligence Review, 58 (7)
  • [45] Role of protein environment on multi-step electron transfer reactions
    Kim, Judy E.
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2011, 242
  • [46] Learning approaches for natural language processing
    Wermter, S.
    Riloff, E.
    Scheler, G.
    Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science), 1040
  • [47] Emergent learning and natural language processing
    Ma, Q
    Lu, BL
    8TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING, VOLS 1-3, PROCEEDING, 2001, : 659 - 664
  • [48] Representation Learning for Natural Language Processing
    刘知远
    林衍凯
    孙茂松
    中文信息学报, 2021, (03) : 143 - 143
  • [49] Deep learning of the natural language processing
    Allauzen, Alexandre
    Schuetze, Hinrich
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2018, 59 (02): : 7 - 14
  • [50] Deep Learning in Natural Language Processing
    Feng, Haoda
    Shi, Feng
    NATURAL LANGUAGE ENGINEERING, 2021, 27 (03) : 373 - 375