Improving Autoregressive Grammatical Error Correction with Non-autoregressive Models

被引:0
|
作者
Cao, Hang [1 ]
Cao, Zhiquan [1 ]
Hu, Chi [1 ]
Hou, Baoyu [1 ]
Xiao, Tong [1 ,2 ]
Zhu, Jingbo [1 ,2 ]
机构
[1] Northeastern Univ, NLP Lab, Sch Comp Sci & Engn, Shenyang, Peoples R China
[2] NiuTrans Res, Shenyang, Peoples R China
基金
美国国家科学基金会; 国家重点研发计划;
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Grammatical Error Correction (GEC) aims to correct grammatical errors in sentences. We find that autoregressive models tend to assign low probabilities to tokens that need corrections. Here we introduce additional signals to the training of GEC models so that these systems can learn to better predict at ambiguous positions. To do this, we use a non-autoregressive model as an auxiliary model, and develop a new regularization term of training by considering the difference in predictions between the autoregressive and non-autoregressive models. We experiment with this method on both English and Chinese GEC tasks. Experimental results show that our GEC system outperforms the baselines on all the data sets significantly.
引用
收藏
页码:12014 / 12027
页数:14
相关论文
共 50 条
  • [21] BANG: Bridging Autoregressive and Non-autoregressive Generation with Large Scale Pretraining
    Qi, Weizhen
    Gong, Yeyun
    Jiao, Jian
    Yan, Yu
    Chen, Weizhu
    Liu, Dayiheng
    Tang, Kewen
    Li, Houqiang
    Chen, Jiusheng
    Zhang, Ruofei
    Zhou, Ming
    Duan, Nan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [22] An Imitation Learning Curriculum for Text Editing with Non-Autoregressive Models
    Agrawal, Sweta
    Carpuat, Marine
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7550 - 7563
  • [23] Partially Non-Autoregressive Image Captioning
    Fei, Zhengcong
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 1309 - 1316
  • [24] Non-autoregressive Error Correction for CTC-based ASR with Phone-conditioned Masked LM
    Futami, Hayato
    Inaguma, Hirofumi
    Ueno, Sei
    Mimura, Masato
    Sakai, Shinsuke
    Kawahara, Tatsuya
    INTERSPEECH 2022, 2022, : 3889 - 3893
  • [25] Non-Autoregressive Transformer for Speech Recognition
    Chen, Nanxin
    Watanabe, Shinji
    Villalba, Jesus
    Zelasko, Piotr
    Dehak, Najim
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 121 - 125
  • [26] Non-Autoregressive Text Generation with Pre-trained Language Models
    Su, Yixuan
    Cai, Deng
    Wang, Yan
    Vandyke, David
    Baker, Simon
    Li, Piji
    Collier, Nigel
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 234 - 243
  • [27] Learning Non-Autoregressive Models from Search for Unsupervised Sentence Summarization
    Liu, Puyuan
    Huang, Chenyang
    Mou, Lili
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7916 - 7929
  • [28] Non-autoregressive Streaming Transformer for Simultaneous Translation
    Ma, Zhengrui
    Zhang, Shaolei
    Guo, Shoutao
    Shao, Chenze
    Zhang, Min
    Feng, Yang
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 5177 - 5190
  • [29] Deep Equilibrium Non-Autoregressive Sequence Learning
    Zheng, Zaixiang
    Zhou, Yi
    Zhou, Hao
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 11763 - 11781
  • [30] Robust Cardinality Estimator by Non-autoregressive Model
    Ito, Ryuichi
    Xiao, Chuan
    Onizuka, Makoto
    SOFTWARE FOUNDATIONS FOR DATA INTEROPERABILITY, SFDI 2021, 2022, 1457 : 55 - 61