Pre-training Intent-Aware Encoders for Zero- and Few-Shot Intent Classification

被引:0
|
作者
Sung, Mujeen [1 ]
Gun, James [2 ]
Mansimov, Elman [2 ]
Pappas, Nikolaos [2 ]
Shu, Raphael [2 ]
Romeo, Salvatore [2 ]
Zhang, Yi [2 ]
Castelli, Vittorio [2 ]
机构
[1] Korea Univ, Seoul, South Korea
[2] AWS AI Labs, New York, NY USA
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Intent classification (IC) plays an important role in task-oriented dialogue systems. However, IC models often generalize poorly when training without sufficient annotated examples for each user intent. We propose a novel pre-training method for text encoders that uses contrastive learning with intent psuedo-labels to produce embeddings that are well-suited for IC tasks, reducing the need for manual annotations. By applying this pre-training strategy, we also introduce Pre-trained Intent-aware Encoder (PIE), which is designed to align encodings of utterances with their intent names. Specifically, we first train a tagger to identify key phrases within utterances that are crucial for interpreting intents. We then use these extracted phrases to create examples for pre-training a text encoder in a contrastive manner. As a result, our PIE model achieves up to 5.4% and 4.0% higher accuracy than the previous state-of-the-art text encoder for the N-way zero- and one-shot settings on four IC datasets.
引用
收藏
页码:10433 / 10442
页数:10
相关论文
共 50 条
  • [21] Diverse Paraphrasing with Insertion Models for Few-Shot Intent Detection
    Chevasson, Raphael
    Laclau, Charlotte
    Gravier, Christophe
    ADVANCES IN INTELLIGENT DATA ANALYSIS XXI, IDA 2023, 2023, 13876 : 65 - 76
  • [22] Semantic Transportation Prototypical Network for Few-shot Intent Detection
    Xu, Weiyuan
    Zhou, Peilin
    You, Chenyu
    Zou, Yuexian
    INTERSPEECH 2021, 2021, : 251 - 255
  • [23] Omni-Training: Bridging Pre-Training and Meta-Training for Few-Shot Learning
    Shu, Yang
    Cao, Zhangjie
    Gao, Jinghan
    Wang, Jianmin
    Yu, Philip S.
    Long, Mingsheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (12) : 15275 - 15291
  • [24] New Intent Discovery with Pre-training and Contrastive Learning
    Zhang, Yuwei
    Zhang, Haode
    Zhan, Li-Ming
    Wu, Xiao-Ming
    Lam, Albert Y. S.
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 256 - 269
  • [25] Self-Supervised Task Augmentation for Few-Shot Intent Detection
    Peng-Fei Sun
    Ya-Wen Ouyang
    Ding-Jie Song
    Xin-Yu Dai
    Journal of Computer Science and Technology, 2022, 37 : 527 - 538
  • [26] Cloze-Style Data Augmentation for Few-Shot Intent Recognition
    Zhang, Xin
    Jiang, Miao
    Chen, Honghui
    Chen, Chonghao
    Zheng, Jianming
    MATHEMATICS, 2022, 10 (18)
  • [27] Few-Shot Intent Detection by Data Augmentation and Class Knowledge Transfer
    Guo, Zhijun
    Niu, Kun
    Chen, Xiao
    Liu, Qi
    Li, Xiao
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 458 - 462
  • [28] EDynamic Semantic Matching and Aggregation Network for Few-shot Intent Detection
    Hoang Nguyen
    Zhang, Chenwei
    Xia, Congying
    Yu, Philip S.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1209 - 1218
  • [29] Boosting few-shot intent detection via feature enrichment and regularization
    Zhang, Feng
    Chen, Wei
    Zhao, Pei
    Wang, Tengjiao
    NEUROCOMPUTING, 2025, 618
  • [30] Two stages prompting for few-shot multi-intent detection
    Zhou, Xingfa
    Yang, Lan
    Wang, Xin
    Zhan, Huayi
    Sun, Rui
    NEUROCOMPUTING, 2024, 579