Hierarchical Multi-Granularity Attention- Based Hybrid Neural Network for Text Classification

被引:6
|
作者
Liu Z. [1 ]
Lu C. [2 ]
Huang H. [2 ]
Lyu S. [2 ]
Tao Z. [3 ,4 ]
机构
[1] School of Information Management for Law, China University of Political Science and Law, Beijing
[2] School of Computer Science, University of Science and Technology of China, Hefei
[3] Division of Life Sciences and Medicine, First Affiliated Hospital of USTC, University of Science and Technology of China, Hefei
[4] Anhui Provincial Cancer Hospital, Hefei
关键词
Attention mechanism; convolutional neural network; multichannel; text classification;
D O I
10.1109/ACCESS.2020.3016727
中图分类号
学科分类号
摘要
Neural network-based approaches have become the driven forces for Natural Language Processing (NLP) tasks. Conventionally, there are two mainstream neural architectures for NLP tasks: the recurrent neural network (RNN) and the convolution neural network (ConvNet). RNNs are good at modeling long-term dependencies over input texts, but preclude parallel computation. ConvNets do not have memory capability and it has to model sequential data as un-ordered features. Therefore, ConvNets fail to learn sequential dependencies over the input texts, but it is able to carry out high-efficient parallel computation. As each neural architecture, such as RNN and ConvNets, has its own pro and con, integration of different architectures is assumed to be able to enrich the semantic representation of texts, thus enhance the performance of NLP tasks. However, few investigation explores the reconciliation of these seemingly incompatible architectures. To address this issue, we propose a hybrid architecture based on a novel hierarchical multi-granularity attention mechanism, named Multi-granularity Attention-based Hybrid Neural Network (MahNN). The attention mechanism is to assign different weights to different parts of the input sequence to increase the computation efficiency and performance of neural models. In MahNN, two types of attentions are introduced: the syntactical attention and the semantical attention. The syntactical attention computes the importance of the syntactic elements (such as words or sentence) at the lower symbolic level and the semantical attention is used to compute the importance of the embedded space dimension corresponding to the upper latent semantics. We adopt the text classification as an exemplifying way to illustrate the ability of MahNN to understand texts. The experimental results on a variety of datasets demonstrate that MahNN outperforms most of the state-of-the-arts for text classification. © 2013 IEEE.
引用
收藏
页码:149362 / 149371
页数:9
相关论文
共 50 条
  • [31] Towards Better Representations for Multi-Label Text Classification with Multi-granularity Information
    Li, Fangfang
    Su, Puzhen
    Duan, Junwen
    Xiao, Weidong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9470 - 9480
  • [32] A multi granularity information fusion text classification model based on attention mechanism
    Chen, Jingfang
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (05) : 7631 - 7645
  • [33] Few-shot learning based on hierarchical classification via multi-granularity relation networks
    Su, Yuling
    Zhao, Hong
    Lin, Yaojin
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2022, 142 : 417 - 429
  • [34] Multi-Granularity Cross-Attention Network for Visual Question Answering
    Wang, Yue
    Gao, Wei
    Cheng, Xinzhou
    Wang, Xin
    Zhao, Huiying
    Xie, Zhipu
    Xu, Lexi
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 2098 - 2103
  • [35] Multi-granularity cross attention network for person re-identification
    Han, Chengmei
    Jiang, Bo
    Tang, Jin
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (10) : 14755 - 14773
  • [36] Multi-granularity cross attention network for person re-identification
    Chengmei Han
    Bo Jiang
    Jin Tang
    Multimedia Tools and Applications, 2023, 82 : 14755 - 14773
  • [37] Information Extraction Network Based on Multi-Granularity Attention and Multi-Scale Self-Learning
    Sun, Weiwei
    Liu, Shengquan
    Liu, Yan
    Kong, Lingqi
    Jian, Zhaorui
    SENSORS, 2023, 23 (09)
  • [38] Multi-granularity Prediction for Scene Text Recognition
    Wang, Peng
    Da, Cheng
    Yao, Cong
    COMPUTER VISION - ECCV 2022, PT XXVIII, 2022, 13688 : 339 - 355
  • [39] Hierarchical Multiple Granularity Attention Network for Long Document Classification
    Hu, Yongli
    Ding, Wen
    Liu, Tengfei
    Gao, Junbin
    Sun, Yanfeng
    Yin, Baocai
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [40] Multi-granularity Separation Network for Text-Based Person Retrieval with Bidirectional Refinement Regularization
    Li, Shenshen
    Xu, Xing
    Shen, Fumin
    Yang, Yang
    PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 307 - 315