Topic-aware hierarchical multi-attention network for text classification

被引:4
|
作者
Jiang, Ye [1 ]
Wang, Yimin [1 ]
机构
[1] Qingdao Univ Sci & Technol, Sch Informat Sci & Technol, Qingdao, Peoples R China
关键词
Text classification; Topic model; Attention mechanism; Natural language processing; LDA;
D O I
10.1007/s13042-022-01734-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks, primarily recurrent and convolutional Neural networks, have been proven successful in text classification. However, convolutional models could be limited when classification tasks are determined by long-range semantic dependency. While the recurrent ones can capture long-range dependency, the sequential architecture of which could constrain the training speed. Meanwhile, traditional networks encode the entire document in a single pass, which omits the hierarchical structure of the document. To address the above issues, this study presents T-HMAN, a Topic-aware Hierarchical Multiple Attention Network for text classification. A multi-head self-attention coupled with convolutional filters is developed to capture long-range dependency via integrating the convolution features from each attention head. Meanwhile, T-HMAN combines topic distributions generated by Latent Dirichlet Allocation (LDA) with sentence-level and document-level inputs respectively in a hierarchical architecture. The proposed model surpasses the accuracies of the current state-of-the-art hierarchical models on five publicly accessible datasets. The ablation study demonstrates that the involvement of multiple attention mechanisms brings significant improvement. The current topic distributions are fixed vectors generated by LDA, the topic distributions will be parameterized and updated simultaneously with the model weights in future work.
引用
收藏
页码:1863 / 1875
页数:13
相关论文
共 50 条
  • [1] Topic-aware hierarchical multi-attention network for text classification
    Ye Jiang
    Yimin Wang
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 1863 - 1875
  • [2] Topic-aware cosine graph convolutional neural network for short text classification
    Min C.
    Chu Y.
    Lin H.
    Wang B.
    Yang L.
    Xu B.
    Soft Computing, 2024, 28 (13-14) : 8119 - 8132
  • [3] Neural Information Diffusion Prediction with Topic-Aware Attention Network
    Wang, Hao
    Yang, Cheng
    Shi, Chuan
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 1899 - 1908
  • [4] TAE: Topic-aware encoder for large-scale multi-label text classification
    Qin, Shaowei
    Wu, Hao
    Zhou, Lihua
    Zhao, Yiji
    Zhang, Lei
    APPLIED INTELLIGENCE, 2024, 54 (08) : 6269 - 6284
  • [5] Hierarchical multi-attention networks for document classification
    Huang, Yingren
    Chen, Jiaojiao
    Zheng, Shaomin
    Xue, Yun
    Hu, Xiaohui
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (06) : 1639 - 1647
  • [6] Enhancing Text Classification by Graph Neural Networks With Multi-Granular Topic-Aware Graph
    Gu, Yongchun
    Wang, Yi
    Zhang, Heng-Ru
    Wu, Jiao
    Gu, Xingquan
    IEEE ACCESS, 2023, 11 : 20169 - 20183
  • [7] Hierarchical Topic-Aware Contextualized Transformers
    Lu, Ruiying
    Chen, Bo
    Guo, Dandan
    Wang, Dongsheng
    Zhou, Mingyuan
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 841 - 852
  • [8] Hierarchical multi-attention networks for document classification
    Yingren Huang
    Jiaojiao Chen
    Shaomin Zheng
    Yun Xue
    Xiaohui Hu
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 1639 - 1647
  • [9] Topic-aware neural attention network for malicious social media spam detection
    Nasser, Maged
    Saeed, Faisal
    Da'u, Aminu
    Alblwi, Abdulaziz
    Al-Sarem, Mohammed
    ALEXANDRIA ENGINEERING JOURNAL, 2025, 111 : 540 - 554
  • [10] Short Text Classification Model Based on Multi-Attention
    Liu, Yunxiang
    Xu, Qi
    2020 13TH INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN (ISCID 2020), 2020, : 225 - 229