Encoding Syntactic Knowledge in Transformer Encoder for Intent Detection and Slot Filling

被引:0
|
作者
Wang, Jixuan [1 ,2 ,3 ]
Wei, Kai [3 ]
Radfar, Martin [3 ]
Zhang, Weiwei [3 ]
Chung, Clement [3 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Vector Inst, Toronto, ON, Canada
[3] Amazon Alexa, Pittsburgh, PA 15205 USA
关键词
NEURAL-NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel Transformer encoder-based architecture with syntactical knowledge encoded for intent detection and slot filling. Specifically, we encode syntactic knowledge into the Transformer encoder by jointly training it to predict syntactic parse ancestors and part-of-speech of each token via multi-task learning. Our model is based on self-attention and feed-forward layers and does not require external syntactic information to be available at inference time. Experiments show that on two benchmark datasets, our models with only two Transformer encoder layers achieve state-of-the-art results. Compared to the previously best performed model without pre-training, our models achieve absolute F1 score and accuracy improvement of 1.59% and 0.85% for slot filling and intent detection on the SNIPS dataset, respectively. Our models also achieve absolute F1 score and accuracy improvement of 0.1% and 0.34% for slot filling and intent detection on the ATIS dataset, respectively, over the previously best performed model. Furthermore, the visualization of the self-attention weights illustrates the benefits of incorporating syntactic information during training.
引用
收藏
页码:13943 / 13951
页数:9
相关论文
共 50 条
  • [21] MISCA: A Joint Model for Multiple Intent Detection and Slot Filling with Intent-Slot Co-Attention
    Pham, Thinh
    Tran, Chi
    Nguyen, Dat Quoc
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12641 - 12650
  • [22] Intent-Slot Correlation Modeling for Joint Intent Prediction and Slot Filling
    Fan, Jun-Feng
    Wang, Mei-Ling
    Li, Chang-Liang
    Zhu, Zi-Qiang
    Mao, Lu
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2022, 37 (02) : 309 - 319
  • [23] Intent-Slot Correlation Modeling for Joint Intent Prediction and Slot Filling
    Jun-Feng Fan
    Mei-Ling Wang
    Chang-Liang Li
    Zi-Qiang Zhu
    Lu Mao
    Journal of Computer Science and Technology, 2022, 37 : 309 - 319
  • [24] Bidirectional Interaction Model for Joint Multiple Intent Detection and Slot Filling
    Li, Shi
    Sun, Zhenpeng
    Computer Engineering and Applications, 2024, 60 (05) : 130 - 138
  • [25] JOINT INTENT DETECTION AND SLOT FILLING BASED ON CONTINUAL LEARNING MODEL
    Hui, Yanfei
    Wang, Jianzong
    Cheng, Ning
    Yu, Fengying
    Wu, Tianbo
    Xiao, Jing
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7643 - 7647
  • [26] A Multi-Task Hierarchical Approach for Intent Detection and Slot Filling
    Firdaus, Mauajama
    Kumar, Ankit
    Ekbal, Asif
    Bhattacharyya, Pushpak
    KNOWLEDGE-BASED SYSTEMS, 2019, 183
  • [27] Multitask learning for multilingual intent detection and slot filling in dialogue systems
    Firdaus, Mauajama
    Ekbal, Asif
    Cambria, Erik
    INFORMATION FUSION, 2023, 91 : 299 - 315
  • [28] A Graph-to-Sequence Model for Joint Intent Detection and Slot Filling
    Wu, Jie
    Harris, Ian G.
    Zhao, Hongzhi
    Ling, Guangming
    2023 IEEE 17TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING, ICSC, 2023, : 131 - 138
  • [29] A Deep Learning Model with Data Enrichment for Intent Detection and Slot Filling
    Dadas, Slawomir
    Protasiewicz, Jaroslaw
    Pedrycz, Witold
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 3012 - 3018
  • [30] Joint intent detection and slot filling for Turkish natural language understanding
    Buyuk, Osman
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2023, 31 (05) : 844 - 859