Label Semantic Aware Pre-training for Few-shot Text Classification

被引:0
|
作者
Mueller, Aaron [1 ]
Krone, Jason [2 ]
Romeo, Salvatore [2 ]
Mansour, Saab [2 ]
Mansimov, Elman [2 ]
Zhang, Yi [2 ]
Roth, Dan [2 ,3 ]
机构
[1] Johns Hopkins Univ, Dept Comp Sci, Baltimore, MD 21218 USA
[2] Amazon Web Serv AI Labs, New York, NY USA
[3] Univ Penn, Dept Comp & Informat Sci, Philadelphia, PA 19104 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In text classification tasks, useful information is encoded in the label names. Label semantic aware systems have leveraged this information for improved text classification performance during fine-tuning and prediction. However, use of label-semantics during pre-training has not been extensively explored. We therefore propose Label Semantic Aware Pre-training (LSAP) to improve the generalization and data efficiency of text classification systems. LSAP incorporates label semantics into pre-trained generative models (T5 in our case) by performing secondary pre-training on labeled sentences from a variety of domains. As domain-general pre-training requires large amounts of data, we develop a filtering and labeling pipeline to automatically create sentence-label pairs from unlabeled text. We perform experiments on intent (ATIS, Snips, TOPv2) and topic classification (AG News, Yahoo! Answers). LSAP obtains significant accuracy improvements over state-of-the-art models for few-shot text classification while maintaining performance comparable to state of the art in high-resource settings.
引用
收藏
页码:8318 / 8334
页数:17
相关论文
共 50 条
  • [1] Effectiveness of Pre-training for Few-shot Intent Classification
    Zhang, Haode
    Zhang, Yuwei
    Zhan, Li-Ming
    Chen, Jiaxin
    Shi, Guangyuan
    Wu, Xiao-Ming
    Lam, Albert Y. S.
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 1114 - 1120
  • [2] Joint Few-Shot Text Classification Aided by Label Semantic and Sentence-Aware Interaction
    Wang, Suhe
    Liu, Bo
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [3] Adversarial training for few-shot text classification
    Croce, Danilo
    Castellucci, Giuseppe
    Basili, Roberto
    [J]. INTELLIGENZA ARTIFICIALE, 2020, 14 (02) : 201 - 214
  • [4] Distinct Label Representations for Few-Shot Text Classification
    Ohashi, Sora
    Takayama, Junya
    Kajiwara, Tomoyuki
    Arase, Yuki
    [J]. ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 831 - 836
  • [5] Uncertainty-aware Self-training for Few-shot Text Classification
    Mukherjee, Subhabrata
    Awadallah, Ahmed Hassan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] Few-Shot Dataset Distillation via Translative Pre-Training
    Liu, Songhua
    Wang, Xinchao
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 18608 - 18618
  • [7] Prompt-Based Label-Aware Framework for Few-Shot Multi-Label Text Classification
    Thaminkaew, Thanakorn
    Lertvittayakumjorn, Piyawat
    Vateekul, Peerapon
    [J]. IEEE ACCESS, 2024, 12 : 28310 - 28322
  • [8] Multitask Pre-training of Modular Prompt for Chinese Few-Shot Learning
    Sun, Tianxiang
    He, Zhengfu
    Zhu, Qin
    Qiu, Xipeng
    Huang, Xuanjing
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 11156 - 11172
  • [9] Synergistic Anchored Contrastive Pre-training for Few-Shot Relation Extraction
    Luo, Da
    Gan, Yanglei
    Hou, Rui
    Lin, Run
    Liu, Qiao
    Cai, Yuxiang
    Gao, Wannian
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18742 - 18750
  • [10] Semantic-Aware Feature Aggregation for Few-Shot Image Classification
    Fusheng Hao
    Fuxiang Wu
    Fengxiang He
    Qieshi Zhang
    Chengqun Song
    Jun Cheng
    [J]. Neural Processing Letters, 2023, 55 : 6595 - 6609