Biomedical word sense disambiguation with bidirectional long short-term memory and attention-based neural networks

被引:13
|
作者
Zhang, Canlin [1 ]
Bis, Daniel [2 ]
Liu, Xiuwen [2 ]
He, Zhe [3 ]
机构
[1] Florida State Univ, Dept Math, Tallahassee, FL 32306 USA
[2] Florida State Univ, Dept Comp Sci, Tallahassee, FL 32306 USA
[3] Florida State Univ, Sch Informat, Tallahassee, FL 32306 USA
关键词
Word sense disambiguation; LSTM; Self-attention; Biomedical;
D O I
10.1186/s12859-019-3079-8
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
Background In recent years, deep learning methods have been applied to many natural language processing tasks to achieve state-of-the-art performance. However, in the biomedical domain, they have not out-performed supervised word sense disambiguation (WSD) methods based on support vector machines or random forests, possibly due to inherent similarities of medical word senses. Results In this paper, we propose two deep-learning-based models for supervised WSD: a model based on bi-directional long short-term memory (BiLSTM) network, and an attention model based on self-attention architecture. Our result shows that the BiLSTM neural network model with a suitable upper layer structure performs even better than the existing state-of-the-art models on the MSH WSD dataset, while our attention model was 3 or 4 times faster than our BiLSTM model with good accuracy. In addition, we trained "universal" models in order to disambiguate all ambiguous words together. That is, we concatenate the embedding of the target ambiguous word to the max-pooled vector in the universal models, acting as a "hint". The result shows that our universal BiLSTM neural network model yielded about 90 percent accuracy. Conclusion Deep contextual models based on sequential information processing methods are able to capture the relative contextual information from pre-trained input word embeddings, in order to provide state-of-the-art results for supervised biomedical WSD tasks.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] Bidirectional long short-term memory attention neural network to estimate neural mass model parameters
    Zhang, Hao
    Yang, Changqing
    Xu, Jingping
    Yuan, Guanli
    Li, Xiaoli
    Gu, Guanghua
    Cui, Dong
    CHAOS SOLITONS & FRACTALS, 2025, 192
  • [42] Bidirectional long short-term memory attention neural network to estimate neural mass model parameters
    Zhang, Hao
    Yang, Changqing
    Xu, Jingping
    Yuan, Guanli
    Li, Xiaoli
    Gu, Guanghua
    Cui, Dong
    Chaos, Solitons and Fractals, 1600,
  • [43] Articulatory Movement Prediction Using Deep Bidirectional Long Short-Term Memory Based Recurrent Neural Networks and Word/Phone Embeddings
    Zhu, Pengcheng
    Xie, Lei
    Chen, Yunlin
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2192 - 2196
  • [44] Relation extraction in Chinese using attention-based bidirectional long short- term networks
    Zhang, Yanzi
    PEERJ COMPUTER SCIENCE, 2023, 9
  • [45] Hybrid Short-term Load Forecasting Method Based on Empirical Wavelet Transform and Bidirectional Long Short-term Memory Neural Networks
    Zhang, Xiaoyu
    Kuenzel, Stefanie
    Colombo, Nicolo
    Watkins, Chris
    JOURNAL OF MODERN POWER SYSTEMS AND CLEAN ENERGY, 2022, 10 (05) : 1216 - 1228
  • [46] VOICE CONVERSION USING DEEP BIDIRECTIONAL LONG SHORT-TERM MEMORY BASED RECURRENT NEURAL NETWORKS
    Sun, Lifa
    Kang, Shiyin
    Li, Kun
    Meng, Helen
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 4869 - 4873
  • [47] Hybrid Short-term Load Forecasting Method Based on Empirical Wavelet Transform and Bidirectional Long Short-term Memory Neural Networks
    Xiaoyu Zhang
    Stefanie Kuenzel
    Nicolo Colombo
    Chris Watkins
    JournalofModernPowerSystemsandCleanEnergy, 2022, 10 (05) : 1216 - 1228
  • [48] BIDIRECTIONAL QUATERNION LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORKS FOR SPEECH RECOGNITION
    Parcollet, Titouan
    Morchid, Mohamed
    Linares, Georges
    De Mori, Renato
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 8519 - 8523
  • [49] Bidirectional Long Short-Term Memory Neural Networks for Linear Sum Assignment Problems
    Minh-Tuan, Nguyen
    Kim, Yong-Hwa
    APPLIED SCIENCES-BASEL, 2019, 9 (17):
  • [50] Influenza surveillance with Baidu index and attention-based long short-term memory model
    Dai, Shangfang
    Han, Litao
    PLOS ONE, 2023, 18 (01):