A Hybrid Neural Network BERT-Cap Based on Pre-Trained Language Model and Capsule Network for User Intent Classification

被引:3
|
作者
Liu, Hai [1 ,2 ]
Liu, Yuanxia [1 ]
Wong, Leung-Pun [3 ]
Lee, Lap-Kei [3 ]
Hao, Tianyong [1 ,4 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou 510000, Peoples R China
[2] Guangzhou Key Lab Big Data & Intelligent Educ, Guangzhou 510000, Peoples R China
[3] Open Univ Hong Kong, Sch Sci & Technol, Kowloon, Hong Kong 999077, Peoples R China
[4] South China Normal Univ, Inst Adv Study Educ Dev Guangdong Hong Kong Macao, Guangzhou 510000, Peoples R China
基金
中国国家自然科学基金;
关键词
Signal encoding - Semantics - Speech processing - Text processing - Encoding (symbols);
D O I
10.1155/2020/8858852
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
User intent classification is a vital component of a question-answering system or a task-based dialogue system. In order to understand the goals of users' questions or discourses, the system categorizes user text into a set of pre-defined user intent categories. User questions or discourses are usually short in length and lack sufficient context; thus, it is difficult to extract deep semantic information from these types of text and the accuracy of user intent classification may be affected. To better identify user intents, this paper proposes a BERT-Cap hybrid neural network model with focal loss for user intent classification to capture user intents in dialogue. The model uses multiple transformer encoder blocks to encode user utterances and initializes encoder parameters with a pre-trained BERT. Then, it extracts essential features using a capsule network with dynamic routing after utterances encoding. Experiment results on four publicly available datasets show that our model BERT-Cap achieves a F1 score of 0.967 and an accuracy of 0.967, outperforming a number of baseline methods, indicating its effectiveness in user intent classification.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] miProBERT: identification of microRNA promoters based on the pre-trained model BERT
    Wang, Xin
    Gao, Xin
    Wang, Guohua
    Li, Dan
    BRIEFINGS IN BIOINFORMATICS, 2023, 24 (03)
  • [32] A Novel Capsule Based Hybrid Neural Network for Sentiment Classification
    Du, Yongping
    Zhao, Xiaozheng
    He, Meng
    Guo, Wenyang
    IEEE ACCESS, 2019, 7 : 39321 - 39328
  • [33] Automatic Topic Labeling model with Paired-Attention based on Pre-trained Deep Neural Network
    He, Dongbin
    Ren, Yanzhao
    Khattak, Abdul Mateen
    Liu, Xinliang
    Tao, Sha
    Gao, Wanlin
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [34] Pre-Trained Deep Convolutional Neural Network for Clostridioides Difficile Bacteria Cytotoxicity Classification Based on Fluorescence Images
    Brodzicki, Andrzej
    Jaworek-Korjakowska, Joanna
    Kleczek, Pawel
    Garland, Megan
    Bogyo, Matthew
    SENSORS, 2020, 20 (23) : 1 - 17
  • [35] A Method of Choosing a Pre-trained Convolutional Neural Network for Transfer Learning in Image Classification Problems
    Trofimov, Alexander G.
    Bogatyreva, Anastasia A.
    ADVANCES IN NEURAL COMPUTATION, MACHINE LEARNING, AND COGNITIVE RESEARCH III, 2020, 856 : 263 - 270
  • [36] Skin Lesion Classification Using Pre-Trained DenseNet201 Deep Neural Network
    Jasil, S. P. Godlin
    Ulagamuthalvi, V.
    ICSPC'21: 2021 3RD INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATION (ICPSC), 2021, : 393 - 396
  • [37] Scanned ECG Arrhythmia Classification Using a Pre-trained Convolutional Neural Network as a Feature Extractor
    Aldosari, Hanadi
    Coenen, Frans
    Lip, Gregory Y. H.
    Zheng, Yalin
    ARTIFICIAL INTELLIGENCE XXXIX, AI 2022, 2022, 13652 : 64 - 80
  • [38] An efficient brain tumor detection and classification using pre-trained convolutional neural network models
    Rao, K. Nishanth
    Khalaf, Osamah Ibrahim
    Krishnasree, V.
    Kumar, Aruru Sai
    Alsekait, Deema Mohammed
    Priyanka, S. Siva
    Alattas, Ahmed Saleh
    AbdElminaam, Diaa Salama
    HELIYON, 2024, 10 (17)
  • [39] On Cognitive Level Classification of Assessment-items Using Pre-trained BERT-based Model
    Dipto, Adnan Saif
    Limon, Md. Mahmudur Rahman
    Tuba, Fatima Tanjum
    Uddin, Md Mohsin
    Khan, M. Saddam Hossain
    Tuhin, Rashedul Amin
    PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND INFORMATION RETRIEVAL, NLPIR 2023, 2023, : 245 - 251
  • [40] CANCN-BERT: A Joint Pre-Trained Language Model for Classical and Modern Chinese
    Ji, Zijing
    Wang, Xin
    Shen, Yuxin
    Rao, Guozheng
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3112 - 3116