Hierarchical and Bidirectional Joint Multi-Task Classifiers for Natural Language Understanding

被引:0
|
作者
Ji, Xiaoyu [1 ,2 ]
Hu, Wanyang [3 ]
Liang, Yanyan [1 ,4 ]
机构
[1] Macau Univ Sci & Technol, Fac Innovat Engn, Sch Comp Sci & Engn, Macau, Peoples R China
[2] Guangxi Key Lab Machine Vis & Intelligent Control, Wuzhou 543002, Peoples R China
[3] Univ Svizzera Italiana, Dept Informat, CH-6962 Lugano, Switzerland
[4] CEI High Tech Res Inst Co Ltd, Macau, Peoples R China
基金
中国国家自然科学基金;
关键词
multi-task classifier; hierarchical structure; bidirectional joint structure; MASSIVE dataset;
D O I
10.3390/math11244895
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
The MASSIVE dataset is a spoken-language comprehension resource package for slot filling, intent classification, and virtual assistant evaluation tasks. It contains multi-language utterances from human beings communicating with a virtual assistant. In this paper, we exploited the relationship between intent classification and slot filling to improve the exact match accuracy by proposing five models with hierarchical and bidirectional architectures. There are two variants for hierarchical architectures and three variants for bidirectional architectures. These are the hierarchical concatenation model, the hierarchical attention-based model, the bidirectional max-pooling model, the bidirectional LSTM model, and the bidirectional attention-based model. The results of our models showed a significant improvement in the averaged exact match accuracy. The hierarchical attention-based model improved the accuracy by 1.01 points for the full training dataset. As for the zero-shot setup, we observed that the exact match accuracy increased from 53.43 to 53.91. In this study, we observed that, for multi-task problems, utilizing the relevance between different tasks can help in improving the model's overall performance.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] Bidirectional Transformer Based Multi-Task Learning for Natural Language Understanding
    Tripathi, Suraj
    Singh, Chirag
    Kumar, Abhay
    Pandey, Chandan
    Jain, Nishant
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2019), 2019, 11608 : 54 - 65
  • [2] A JOINT MULTI-TASK LEARNING FRAMEWORK FOR SPOKEN LANGUAGE UNDERSTANDING
    Li, Changliang
    Kong, Cunliang
    Zhao, Yan
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 6054 - 6058
  • [3] Multi-Task Deep Neural Networks for Natural Language Understanding
    Liu, Xiaodong
    He, Pengcheng
    Chen, Weizhu
    Gao, Jianfeng
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 4487 - 4496
  • [4] Hierarchical Multi-Task Natural Language Understanding for Cross-domain Conversational AI: HERMIT NLU
    Vanzo, Andrea
    Bastianelli, Emanuele
    Lemon, Oliver
    20TH ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2019), 2019, : 254 - 263
  • [5] Multi-task learning for Joint Language Understanding and Dialogue State Tracking
    Rastogi, Abhinav
    Gupta, Raghav
    Hakkani-Tur, Dilek
    19TH ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2018), 2018, : 376 - 384
  • [6] BAM! Born-Again Multi-Task Networks for Natural Language Understanding
    Clark, Kevin
    Minh-Thang Luong
    Khandelwal, Urvashi
    Manning, Christopher D.
    Le, Quoc V.
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5931 - 5937
  • [7] The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural Language Understanding
    Liu, Xiaodong
    Wang, Yu
    Ji, Jianshu
    Cheng, Hao
    Zhu, Xueyun
    Awa, Emmanuel
    He, Pengcheng
    Chen, Weizhu
    Poon, Hoifung
    Cao, Guihong
    Gao, Jianfeng
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020): SYSTEM DEMONSTRATIONS, 2020, : 118 - 126
  • [8] Multi-task Learning with Bidirectional Language Models for Text Classification
    Yang, Qi
    Shang, Lin
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [9] Macular: a Multi-Task Adversarial Framework for Cross-Lingual Natural Language Understanding
    Wang, Haoyu
    Wang, Yaqing
    Wu, Feijie
    Xue, Hongfei
    Gao, Jing
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5061 - 5070
  • [10] Multi-task learning approach for utilizing temporal relations in natural language understanding tasks
    Lim, Chae-Gyun
    Jeong, Young-Seob
    Choi, Ho-Jin
    SCIENTIFIC REPORTS, 2023, 13 (01)