Bidirectional LSTM with attention mechanism and convolutional layer for text classification

被引:669
|
作者
Liu, Gang [1 ]
Guo, Jiabao [1 ]
机构
[1] Hubei Univ Technol, Sch Comp Sci, Wuhan 430072, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Long short-term memory; Attention mechanism; Natural language processing; Text classification; SHORT-TERM-MEMORY; NEURAL-NETWORKS;
D O I
10.1016/j.neucom.2019.01.078
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural network models have been widely used in the field of natural language processing (NLP). Recurrent neural networks (RNNs), which have the ability to process sequences of arbitrary length, are common methods for sequence modeling tasks. Long short-term memory (LSTM) is one kind of RNNs and has achieved remarkable performance in text classification. However, due to the high dimensionality and sparsity of text data, and to the complex semantics of the natural language, text classification presents difficult challenges. In order to solve the above problems, a novel and unified architecture which contains a bidirectional LSTM (BiLSTM), attention mechanism and the convolutional layer is proposed in this paper. The proposed architecture is called attention-based bidirectional long short-term memory with convolution layer (AC-BiLSTM). In AC-BiLSTM, the convolutional layer extracts the higher-level phrase representations from the word embedding vectors and BiLSTM is used to access both the preceding and succeeding context representations. Attention mechanism is employed to give different focus to the information out-putted from the hidden layers of BiLSTM. Finally, the softmax classifier is used to classify the processed context information. AC-BiLSTM is able to capture both the local feature of phrases as well as global sentence semantics. Experimental verifications are conducted on six sentiment classification datasets and a question classification dataset, including detailed analysis for AC-BiLSTM. The results clearly show that AC-BiLSTM outperforms other state-of-the-art text classification methods in terms of the classification accuracy. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:325 / 338
页数:14
相关论文
共 50 条
  • [11] Multi-channel Attention Mechanism Text Classification Model Based on CNN and LSTM
    Teng, Jinbao
    Kong, Weiwei
    Tian, Qiaoxin
    Wang, Zhaoqian
    Li, Long
    Computer Engineering and Applications, 2024, 57 (23) : 154 - 162
  • [12] Bidirectional LSTM with self-attention mechanism and multi-channel features for sentiment classification
    Li, Weijiang
    Qi, Fang
    Tang, Ming
    Yu, Zhengtao
    NEUROCOMPUTING, 2020, 387 : 63 - 77
  • [13] Hierarchical Convolutional Attention Networks for Text Classification
    Gao, Shang
    Ramanathan, Arvind
    Tourassi, Georgia
    REPRESENTATION LEARNING FOR NLP, 2018, : 11 - 23
  • [14] Multi-Label Text Classification Combining Bidirectional Attention and Contrast Enhancement Mechanism
    Li, Jiandong
    Fu, Jia
    Li, Jiaqi
    Computer Engineering and Applications, 2024, 60 (16) : 105 - 115
  • [15] Bidirectional Gated Temporal Convolution with Attention for text classification
    Ren, Jiansi
    Wu, Wei
    Liu, Gang
    Chen, Zhe
    Wang, Ruoxiang
    NEUROCOMPUTING, 2021, 455 : 265 - 273
  • [16] Outpatient Text Classification Using Attention-Based Bidirectional LSTM for Robot-Assisted Servicing in Hospital
    Chen, Che-Wen
    Tseng, Shih-Pang
    Kuan, Ta-Wen
    Wang, Jhing-Fa
    INFORMATION, 2020, 11 (02)
  • [17] AB-LSTM: Attention-based Bidirectional LSTM Model for Scene Text Detection
    Liu, Zhandong
    Zhou, Wengang
    Li, Houqiang
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2019, 15 (04)
  • [18] Vehicle Destination Prediction Using Bidirectional LSTM with Attention Mechanism
    Casabianca, Pietro
    Zhang, Yu
    Martinez-Garcia, Miguel
    Wan, Jiafu
    SENSORS, 2021, 21 (24)
  • [19] Novel double-layer bidirectional LSTM network with improved attention mechanism for predicting energy consumption
    He, Yan-Lin
    Chen, Lei
    Gao, Yanlu
    Ma, Jia-Hui
    Xu, Yuan
    Zhu, Qun-Xiong
    ISA TRANSACTIONS, 2022, 127 : 350 - 360
  • [20] A Self-attention Based LSTM Network for Text Classification
    Jing, Ran
    2019 3RD INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2019), 2019, 1207