An Imitation Learning Curriculum for Text Editing with Non-Autoregressive Models

被引:0
|
作者
Agrawal, Sweta [1 ]
Carpuat, Marine [1 ]
机构
[1] Univ Maryland, Dept Comp Sci, College Pk, MD 20742 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a framework for training non-autoregressive sequence-to-sequence models for editing tasks, where the original input sequence is iteratively edited to produce the output. We show that the imitation learning algorithms designed to train such models for machine translation introduces mismatches between training and inference that lead to undertraining and poor generalization in editing scenarios. We address this issue with two complementary strategies: 1) a roll-in policy that exposes the model to intermediate training sequences that it is more likely to encounter during inference, 2) a curriculum that presents easy-to-learn edit operations first, gradually increasing the difficulty of training samples as the model becomes competent. We show the efficacy of these strategies on two challenging English editing tasks: controllable text simplification and abstractive summarization. Our approach significantly improves output quality on both tasks and controls output complexity better on the simplification task.
引用
收藏
页码:7550 / 7563
页数:14
相关论文
共 50 条
  • [41] Pushing the Limits of Non-Autoregressive Speech Recognition
    Ng, Edwin G.
    Chiu, Chung-Cheng
    Zhang, Yu
    Chan, William
    [J]. INTERSPEECH 2021, 2021, : 3725 - 3729
  • [42] Acyclic Transformer for Non-Autoregressive Machine Translation
    Huang, Fei
    Zhou, Hao
    Liu, Yang
    Li, Hang
    Huang, Minlie
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [43] Non-Autoregressive Machine Translation with Auxiliary Regularization
    Wang, Yiren
    Tian, Fei
    He, Di
    Qin, Tao
    Zhai, ChengXiang
    Liu, Tie-Yan
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 5377 - 5384
  • [44] Knowledge Transfer and Distillation from Autoregressive to Non-Autoregressive Speech Recognition
    Gong, Xun
    Zhou, Zhikai
    Qian, Yanmin
    [J]. INTERSPEECH 2022, 2022, : 2618 - 2622
  • [45] BANG: Bridging Autoregressive and Non-autoregressive Generation with Large Scale Pretraining
    Qi, Weizhen
    Gong, Yeyun
    Jiao, Jian
    Yan, Yu
    Chen, Weizhu
    Liu, Dayiheng
    Tang, Kewen
    Li, Houqiang
    Chen, Jiusheng
    Zhang, Ruofei
    Zhou, Ming
    Duan, Nan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [46] Non-Autoregressive Machine Translation with Latent Alignments
    Saharia, Chitwan
    Chan, William
    Saxena, Saurabh
    Norouzi, Mohammad
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1098 - 1108
  • [47] A Survey of Non-Autoregressive Neural Machine Translation
    Li, Feng
    Chen, Jingxian
    Zhang, Xuejun
    [J]. ELECTRONICS, 2023, 12 (13)
  • [48] AligNART: Non-autoregressive Neural Machine Translation by Jointly Learning to Estimate Alignment and Translate
    Song, Jongyoon
    Kim, Sungwon
    Yoon, Sungroh
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1 - 14
  • [49] Non-Autoregressive Image Captioning with Counterfactuals-Critical Multi-Agent Learning
    Guo, Longteng
    Liu, Jing
    Zhu, Xinxin
    He, Xingjian
    Jiang, Jie
    Lu, Hanging
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 767 - 773
  • [50] Hierarchical and Multi-Scale Variational Autoencoder for Diverse and Natural Non-Autoregressive Text-to-Speech
    Bae, Jae-Sung
    Yang, Jinhyeok
    Bak, Tae-Jun
    Joo, Young-Sun
    [J]. INTERSPEECH 2022, 2022, : 813 - 817