Efficient Semi-supervised Consistency Training for Natural Language Understanding

被引:0
|
作者
Leung, George [1 ]
Tan, Joshua [1 ]
机构
[1] Amazon Alexa AI, Sacramento, CA 95833 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Manually labeled training data is expensive, noisy, and often scarce, such as when developing new features or localizing existing features for a new region. In cases where labeled data is limited but unlabeled data is abundant, semi-supervised learning methods such as consistency training can be used to improve model performance, by training models to output consistent predictions between original and augmented versions of unlabeled data. In this work, we explore different data augmentation methods for consistency training (CT) on Natural Language Understanding (NLU) domain classification (DC) in the limited labeled-data regime. We explore three types of augmentation techniques (human paraphrasing, back-translation, and dropout) for unlabeled data and train DC models to jointly minimize both the supervised loss and the consistency loss on unlabeled data. Our results demonstrate that DC models trained with CT methods and dropout-based augmentation on only 0.1% (2,998 instances) of labeled data with the remainder as unlabeled can achieve a top-1 relative accuracy reduction of 12.25% compared to fully supervised model trained with 100% of labeled data, outperforming fully supervised models trained on 10x that amount of labeled data. The dropout-based augmentation achieves similar performance compare to back-translation-based augmentation with much less computational resources. This paves the way for applications of using large scale unlabeled data for semi-supervised learning in production NLU systems.
引用
收藏
页码:86 / 93
页数:8
相关论文
共 50 条
  • [1] EFFICIENT SEMI-SUPERVISED LEARNING FOR NATURAL LANGUAGE UNDERSTANDING BY OPTIMIZING DIVERSITY
    Cho, Eunah
    Xie, He
    Lalor, John P.
    Kumar, Varun
    Campbell, William M.
    [J]. 2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 1077 - 1084
  • [2] Dual Learning for Semi-Supervised Natural Language Understanding
    Zhu, Su
    Cao, Ruisheng
    Yu, Kai
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 1936 - 1947
  • [3] Semi-Supervised Learning of Statistical Models for Natural Language Understanding
    Zhou, Deyu
    He, Yulan
    [J]. SCIENTIFIC WORLD JOURNAL, 2014,
  • [4] Industry Scale Semi-Supervised Learning for Natural Language Understanding
    Chen, Luoxin
    Garcia, Francisco
    Kumar, Varun
    Xie, He
    Lu, Jianhua
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2021, 2021, : 311 - 318
  • [5] Interpolation Consistency Training for Semi-Supervised Learning
    Verma, Vikas
    Lamb, Alex
    Kannala, Juho
    Bengio, Yoshua
    Lopez-Paz, David
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3635 - 3641
  • [6] Interpolation consistency training for semi-supervised learning
    Verma, Vikas
    Kawaguchi, Kenji
    Lamb, Alex
    Kannala, Juho
    Solin, Arno
    Bengio, Yoshua
    Lopez-Paz, David
    [J]. NEURAL NETWORKS, 2022, 145 : 90 - 106
  • [7] Semi-Supervised Formality Style Transfer with Consistency Training
    Liu, Ao
    Wang, An
    Okazaki, Naoaki
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4689 - 4701
  • [8] CGT: Consistency Guided Training in Semi-Supervised Learning
    Hasan, Nesreen
    Ghorban, Farzin
    Velten, Joerg
    Kummert, Anton
    [J]. PROCEEDINGS OF THE 17TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISAPP), VOL 5, 2022, : 55 - 64
  • [9] A Semi-supervised Method for Efficient Construction of Statistical Spoken Language Understanding Resources
    Kim, Seokhwan
    Jeong, Minwoo
    Lee, Gary Geunbae
    [J]. INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 977 - 980
  • [10] Semi-supervised Left Atrium Segmentation with Mutual Consistency Training
    Wu, Yicheng
    Xu, Minfeng
    Ge, Zongyuan
    Cai, Jianfei
    Zhang, Lei
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT II, 2021, 12902 : 297 - 306