A Hybrid Neural Network BERT-Cap Based on Pre-Trained Language Model and Capsule Network for User Intent Classification

被引:3
|
作者
Liu, Hai [1 ,2 ]
Liu, Yuanxia [1 ]
Wong, Leung-Pun [3 ]
Lee, Lap-Kei [3 ]
Hao, Tianyong [1 ,4 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou 510000, Peoples R China
[2] Guangzhou Key Lab Big Data & Intelligent Educ, Guangzhou 510000, Peoples R China
[3] Open Univ Hong Kong, Sch Sci & Technol, Kowloon, Hong Kong 999077, Peoples R China
[4] South China Normal Univ, Inst Adv Study Educ Dev Guangdong Hong Kong Macao, Guangzhou 510000, Peoples R China
基金
中国国家自然科学基金;
关键词
Signal encoding - Semantics - Speech processing - Text processing - Encoding (symbols);
D O I
10.1155/2020/8858852
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
User intent classification is a vital component of a question-answering system or a task-based dialogue system. In order to understand the goals of users' questions or discourses, the system categorizes user text into a set of pre-defined user intent categories. User questions or discourses are usually short in length and lack sufficient context; thus, it is difficult to extract deep semantic information from these types of text and the accuracy of user intent classification may be affected. To better identify user intents, this paper proposes a BERT-Cap hybrid neural network model with focal loss for user intent classification to capture user intents in dialogue. The model uses multiple transformer encoder blocks to encode user utterances and initializes encoder parameters with a pre-trained BERT. Then, it extracts essential features using a capsule network with dynamic routing after utterances encoding. Experiment results on four publicly available datasets show that our model BERT-Cap achieves a F1 score of 0.967 and an accuracy of 0.967, outperforming a number of baseline methods, indicating its effectiveness in user intent classification.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Automated Classification of Urinary Cells: Using Convolutional Neural Network Pre-trained on Lung Cells
    Teramoto, Atsushi
    Michiba, Ayano
    Kiriyama, Yuka
    Sakurai, Eiko
    Shiroki, Ryoichi
    Tsukamoto, Tetsuya
    APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [42] Drug-BERT : Pre-trained Language Model Specialized for Korean Drug Crime
    Lee, Jeong Min
    Lee, Suyeon
    Byon, Sungwon
    Jung, Eui-Suk
    Baek, Myung-Sun
    19TH IEEE INTERNATIONAL SYMPOSIUM ON BROADBAND MULTIMEDIA SYSTEMS AND BROADCASTING, BMSB 2024, 2024, : 186 - 188
  • [43] A Filter for SAR Image Despeckling Using Pre-Trained Convolutional Neural Network Model
    Pan, Ting
    Peng, Dong
    Yang, Wen
    Li, Heng-Chao
    REMOTE SENSING, 2019, 11 (20)
  • [44] Pre-trained Deep Convolution Neural Network Model With Attention for Speech Emotion Recognition
    Zhang, Hua
    Gou, Ruoyun
    Shang, Jili
    Shen, Fangyao
    Wu, Yifan
    Dai, Guojun
    FRONTIERS IN PHYSIOLOGY, 2021, 12
  • [45] BERT-siRNA: siRNA target prediction based on BERT pre-trained interpretable model
    Xu, Jiayu
    Xu, Nan
    Xie, Weixin
    Zhao, Chengkui
    Yu, Lei
    Feng, Weixing
    GENE, 2024, 910
  • [46] Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks
    Malsagov, M. Yu
    Khayrov, E. M.
    Pushkareva, M. M.
    Karandashev, I. M.
    OPTICAL MEMORY AND NEURAL NETWORKS, 2019, 28 (04) : 262 - 270
  • [47] Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks
    M. Yu. Malsagov
    E. M. Khayrov
    M. M. Pushkareva
    I. M. Karandashev
    Optical Memory and Neural Networks, 2019, 28 : 262 - 270
  • [48] BERT-Based Hybrid RNN Model for Multi-class Text Classification to Study the Effect of Pre-trained Word Embeddings
    Shreyashree, S.
    Sunagar, Pramod
    Rajarajeswari, S.
    Kanavalli, Anita
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (09) : 667 - 674
  • [49] Enriching Pre-trained Language Model with Entity Information for Relation Classification
    Wu, Shanchan
    He, Yifan
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2361 - 2364
  • [50] Pre-Trained Joint Model for Intent Classification and Slot Filling with Semantic Feature Fusion
    Chen, Yan
    Luo, Zhenghang
    SENSORS, 2023, 23 (05)