Pre-training Intent-Aware Encoders for Zero- and Few-Shot Intent Classification

被引:0
|
作者
Sung, Mujeen [1 ]
Gun, James [2 ]
Mansimov, Elman [2 ]
Pappas, Nikolaos [2 ]
Shu, Raphael [2 ]
Romeo, Salvatore [2 ]
Zhang, Yi [2 ]
Castelli, Vittorio [2 ]
机构
[1] Korea Univ, Seoul, South Korea
[2] AWS AI Labs, New York, NY USA
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Intent classification (IC) plays an important role in task-oriented dialogue systems. However, IC models often generalize poorly when training without sufficient annotated examples for each user intent. We propose a novel pre-training method for text encoders that uses contrastive learning with intent psuedo-labels to produce embeddings that are well-suited for IC tasks, reducing the need for manual annotations. By applying this pre-training strategy, we also introduce Pre-trained Intent-aware Encoder (PIE), which is designed to align encodings of utterances with their intent names. Specifically, we first train a tagger to identify key phrases within utterances that are crucial for interpreting intents. We then use these extracted phrases to create examples for pre-training a text encoder in a contrastive manner. As a result, our PIE model achieves up to 5.4% and 4.0% higher accuracy than the previous state-of-the-art text encoder for the N-way zero- and one-shot settings on four IC datasets.
引用
收藏
页码:10433 / 10442
页数:10
相关论文
共 50 条
  • [31] Self-Supervised Task Augmentation for Few-Shot Intent Detection
    Sun, Peng-Fei
    Ouyang, Ya-Wen
    Song, Ding-Jie
    Dai, Xin-Yu
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2022, 37 (03) : 527 - 538
  • [32] Enhancing cross-encoders using knowledge graph hierarchy for medical entity linking in zero- and few-shot scenarios
    Gallego, Fernando
    Ruas, Pedro
    Couto, Francisco M.
    Veredas, Francisco J.
    KNOWLEDGE-BASED SYSTEMS, 2025, 314
  • [33] Does Utterance entails Intent?: Evaluating Natural Language Inference Based Setup for Few-Shot Intent Detection
    Kumar, Ayush
    Malik, Vijit
    Vepa, Jithendra
    INTERSPEECH 2022, 2022, : 4501 - 4505
  • [34] Label Agnostic Pre-training for Zero-shot Text Classification
    Clarke, Christopher
    Heng, Yuzhao
    Kang, Yiping
    Flautner, Krisztian
    Tang, Lingjia
    Mars, Jason
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1009 - 1021
  • [35] Automatic Text Classification With Large Language Models: A Review of <monospace>openai</monospace> for Zero- and Few-Shot Classification
    Anglin, Kylie L.
    Ventura, Claudia
    JOURNAL OF EDUCATIONAL AND BEHAVIORAL STATISTICS, 2024,
  • [36] Label Verbalization and Entailment for Effective Zero- and Few-Shot Relation Extraction
    Sainz, Oscar
    de Lacalle, Oier Lopez
    Labaka, Gorka
    Barrena, Ander
    Agirre, Eneko
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1199 - 1212
  • [37] Generalized Zero- and Few-Shot Learning via Aligned Variational Autoencoders
    Schonfeld, Edgar
    Ebrahimi, Sayna
    Sinha, Samarth
    Darrell, Trevor
    Akata, Zeynep
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 8239 - 8247
  • [38] LEARNING DISCRIMINATIVE LATENT FEATURES FOR GENERALIZED ZERO- AND FEW-SHOT LEARNING
    Huang, Yijie
    Deng, Zhenghong
    Wu, Tao
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [39] TabPrompt: Graph-based Pre-training and Prompting for Few-shot Table Understanding
    Jin, Rihui
    Wang, Jianan
    Tan, Wei
    Chen, YongRui
    Qi, Guilin
    Hao, Wang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 7373 - 7383
  • [40] Few-Shot Contrastive Learning-Based Multi-Round Dialogue Intent Classification Method
    Wei, Feng
    Zhang, Xu
    EXPERT SYSTEMS, 2025, 42 (02)