Model-Agnostic Zero-Shot Intent Detection via Contrastive Transfer Learning

被引:2
|
作者
Maqbool, M. H. [1 ]
Fereidouni, Moghis [2 ]
Siddique, A. B. [2 ]
Foroosh, Hassan [1 ]
机构
[1] Univ Cent Florida, Orlando, FL USA
[2] Univ Kentucky, Lexington, KY 40506 USA
关键词
Intent detection; zero-shot learning; dialog systems;
D O I
10.1142/S1793351X24410010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An intent detector is a central component of any task-oriented conversational system. The goal of the intent detector is to identify the user's goal by classifying natural language utterances. In recent years, research has focused on supervised intent detection models. Supervised learning approaches cannot accommodate unseen intents, which may emerge after the system has been deployed- the more practically relevant setting, known as zero-shot intent detection. The existing zero-shot learning approaches split a dataset into seen and unseen intents for training and evaluations without taking the sensitivity of the data collection process into account. That is, humans tend to use repeated vocabulary and compose sentences with similar compositional structures. We argue that the source-to-target relationship learning objective of zero-shot approaches under typical data split procedure renders the zero-shot models prone to misclassifications when target intents are divergent from source intents. To this end, we propose INTEND, a zero-shot INTENt Detection methodology that leverages contrastive transfer learning and employs a zero-shot learning paradigm in its true sense. First, in contrast to partitioning the training and testing sets from the same dataset, we demonstrate that selecting training and testing sets from two different datasets allows for rigorous zero-shot intent detection evaluations. Second, our employed contrastive learning goal is model-agnostic and encourages the system to focus on learning a generic similarity function, rather than on commonly encountered patterns in the training set. We conduct extensive experimental evaluations using a range of transformer models on four public intent detection datasets for up to 150 unseen classes. Our experimental results show that INTEND consistently outperforms state-of-the-art zero-shot techniques by a substantial margin. Furthermore, our approach achieves significantly better performance than few-shot intent detection models.
引用
收藏
页码:5 / 24
页数:20
相关论文
共 50 条
  • [31] Semantics Guided Contrastive Learning of Transformers for Zero-shot Temporal Activity Detection
    Nag, Sayak
    Goldstein, Orpaz
    Roy-Chowdhury, Amit K.
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 6232 - 6242
  • [32] Unknown Intent Detection Using Gaussian Mixture Model with an Application to Zero-shot Intent Classification
    Yan, Guangfeng
    Fan, Lu
    Li, Qimai
    Liu, Han
    Zhang, Xiaotong
    Wu, Xiao-Ming
    Lam, Albert Y. S.
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 1050 - 1060
  • [33] Zero-Shot-BERT-Adapters: a Zero-Shot Pipeline for Unknown Intent Detection
    Comi, Daniele
    Christofidellis, Dimitrios
    Piazza, Pier Francesco
    Manica, Matteo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 650 - 663
  • [34] Meta-learning with topic-agnostic representations for zero-shot stance detection
    Choi, Kyuri
    Ko, Youngjoong
    PATTERN RECOGNITION LETTERS, 2023, 171 : 15 - 20
  • [35] Zero-Shot Learning for Intrusion Detection via Attribute Representation
    Li, Zhipeng
    Qin, Zheng
    Shen, Pengbo
    Jiang, Liu
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 352 - 364
  • [36] Zero Shot on the Cold-Start Problem: Model-Agnostic Interest Learning for Recommender Systems
    Feng, Philip J.
    Pan, Pingjun
    Zhou, Tingting
    Chen, Hongxiang
    Luo, Chuanjiang
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 474 - 483
  • [37] Zero-shot stance detection via multi-perspective contrastive with unlabeled data
    Jiang, Yan
    Gao, Jinhua
    Shen, Huawei
    Cheng, Xueqi
    INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (04)
  • [38] A meta-contrastive learning with data augmentation framework for zero-shot stance detection
    Wang, Chunling
    Zhang, Yijia
    Wang, Shilong
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 250
  • [39] Unleashing the Power of Contrastive Learning for Zero-Shot Video Summarization
    Pang, Zongshang
    Nakashima, Yuta
    Otani, Mayu
    Nagahara, Hajime
    JOURNAL OF IMAGING, 2024, 10 (09)
  • [40] Dual Prototype Contrastive Network for Generalized Zero-Shot Learning
    Jiang, Huajie
    Li, Zhengxian
    Hu, Yongli
    Yin, Baocai
    Yang, Jian
    van den Hengel, Anton
    Yang, Ming-Hsuan
    Qi, Yuankai
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1111 - 1122