Model-Agnostic Zero-Shot Intent Detection via Contrastive Transfer Learning

被引:2
|
作者
Maqbool, M. H. [1 ]
Fereidouni, Moghis [2 ]
Siddique, A. B. [2 ]
Foroosh, Hassan [1 ]
机构
[1] Univ Cent Florida, Orlando, FL USA
[2] Univ Kentucky, Lexington, KY 40506 USA
关键词
Intent detection; zero-shot learning; dialog systems;
D O I
10.1142/S1793351X24410010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An intent detector is a central component of any task-oriented conversational system. The goal of the intent detector is to identify the user's goal by classifying natural language utterances. In recent years, research has focused on supervised intent detection models. Supervised learning approaches cannot accommodate unseen intents, which may emerge after the system has been deployed- the more practically relevant setting, known as zero-shot intent detection. The existing zero-shot learning approaches split a dataset into seen and unseen intents for training and evaluations without taking the sensitivity of the data collection process into account. That is, humans tend to use repeated vocabulary and compose sentences with similar compositional structures. We argue that the source-to-target relationship learning objective of zero-shot approaches under typical data split procedure renders the zero-shot models prone to misclassifications when target intents are divergent from source intents. To this end, we propose INTEND, a zero-shot INTENt Detection methodology that leverages contrastive transfer learning and employs a zero-shot learning paradigm in its true sense. First, in contrast to partitioning the training and testing sets from the same dataset, we demonstrate that selecting training and testing sets from two different datasets allows for rigorous zero-shot intent detection evaluations. Second, our employed contrastive learning goal is model-agnostic and encourages the system to focus on learning a generic similarity function, rather than on commonly encountered patterns in the training set. We conduct extensive experimental evaluations using a range of transformer models on four public intent detection datasets for up to 150 unseen classes. Our experimental results show that INTEND consistently outperforms state-of-the-art zero-shot techniques by a substantial margin. Furthermore, our approach achieves significantly better performance than few-shot intent detection models.
引用
收藏
页码:5 / 24
页数:20
相关论文
共 50 条
  • [41] Contrastive visual feature filtering for generalized zero-shot learning
    Meng, Shixuan
    Jiang, Rongxin
    Tian, Xiang
    Zhou, Fan
    Chen, Yaowu
    Liu, Junjie
    Shen, Chen
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024,
  • [42] Siamese Contrastive Embedding Network for Compositional Zero-Shot Learning
    Li, Xiangyu
    Yang, Xu
    Wei, Kun
    Deng, Cheng
    Yang, Muli
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9316 - 9325
  • [43] Zero-shot object detection with contrastive semantic association network
    Li, Haohe
    Wang, Chong
    Liu, Weijie
    Gong, Yilin
    Dai, Xinmiao
    APPLIED INTELLIGENCE, 2023, 53 (24) : 30056 - 30068
  • [44] Zero-shot object detection with contrastive semantic association network
    Haohe Li
    Chong Wang
    Weijie Liu
    Yilin Gong
    Xinmiao Dai
    Applied Intelligence, 2023, 53 : 30056 - 30068
  • [45] Zero-shot Event Extraction via Transfer Learning: Challenges and Insights
    Lyu, Qing
    Zhang, Hongming
    Sulem, Elior
    Roth, Dan
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 322 - 332
  • [46] Zero-Shot Transfer Learning of a Throwing Task via Domain Randomization
    Park, Sungyong
    Kim, Jigang
    Kim, H. Jin
    2020 20TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2020, : 1026 - 1030
  • [47] CONNECTING TARGETS VIA LATENT TOPICS AND CONTRASTIVE LEARNING: A UNIFIED FRAMEWORK FOR ROBUST ZERO-SHOT AND FEW-SHOT STANCE DETECTION
    Liu, Rui
    Lin, Zheng
    Fu, Peng
    Liu, Yuanxin
    Wang, Weiping
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7812 - 7816
  • [48] Relational Knowledge Transfer for Zero-Shot Learning
    Wang, Donghui
    Li, Yanan
    Lin, Yuetan
    Zhuang, Yueting
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2145 - 2151
  • [49] Multi-granularity contrastive zero-shot learning model based on attribute decomposition
    Wang, Yuanlong
    Wang, Jing
    Fan, Yue
    Chai, Qinghua
    Zhang, Hu
    Li, Xiaoli
    Li, Ru
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (01)
  • [50] Hypernetworks for Zero-Shot Transfer in Reinforcement Learning
    Rezaei-Shoshtari, Sahand
    Morissette, Charlotte
    Hogan, Francois R.
    Dudek, Gregory
    Meger, David
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9579 - 9587