Restricted or Not: A General Training Framework for Neural Machine Translation

被引:0
|
作者
Li, Zuchao [1 ,2 ]
Utiyama, Masao [3 ]
Sumita, Eiichiro [3 ]
Zhao, Hai [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai, Peoples R China
[2] Shanghai Jiao Tong Univ, AI Inst, MoE Key Lab Artificial Intelligence, Shanghai, Peoples R China
[3] Natl Inst Informat & Commun Technol NICT, Kyoto, Japan
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Restricted machine translation incorporates human prior knowledge into translation. It restricts the flexibility of the translation to satisfy the demands of translation in specific scenarios. Existing work typically imposes constraints on beam search decoding. Although this can satisfy the requirements overall, it usually requires a larger beam size and far longer decoding time than unrestricted translation, which limits the concurrent processing ability of the translation model in deployment, and thus its practicality. In this paper, we propose a general training framework that allows a model to simultaneously support both unrestricted and restricted translation by adopting an additional auxiliary training process without constraining the decoding process. This maintains the benefits of restricted translation but greatly reduces the extra time overhead of constrained decoding, thus improving its practicality. The effectiveness of our proposed training framework is demonstrated by experiments on both original (WAT21 En <-> Ja) and simulated (WMT14 En -> De and En -> Fr) restricted translation benchmarks.
引用
收藏
页码:245 / 251
页数:7
相关论文
共 50 条
  • [1] A General Framework for Adaptation of Neural Machine Translation to Simultaneous Translation
    Chen, Yun
    Li, Liangyou
    Jiang, Xin
    Chen, Xiao
    Liu, Qun
    1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 191 - 200
  • [2] ZeUS: An Unified Training Framework for Constrained Neural Machine Translation
    Yang, Murun
    IEEE ACCESS, 2024, 12 : 124695 - 124704
  • [3] Confidence Based Bidirectional Global Context Aware Training Framework for Neural Machine Translation
    Zhou, Chulun
    Meng, Fandong
    Zhou, Jie
    Zhang, Min
    Wang, Hongji
    Su, Jinsong
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2878 - 2889
  • [4] Improving Neural Machine Translation by Bidirectional Training
    Ding, Liang
    Wu, Di
    Tao, Dacheng
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3278 - 3284
  • [5] Discriminant training of neural networks for machine translation
    Quoc-Khanh Do
    Allauzen, Alexandre
    Yvon, Francois
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2016, 57 (01): : 111 - 135
  • [6] Generative adversarial training for neural machine translation
    Yang, Zhen
    Chen, Wei
    Wang, Feng
    Xu, Bo
    NEUROCOMPUTING, 2018, 321 : 146 - 155
  • [7] Speed Up the Training of Neural Machine Translation
    Liu, Xinyue
    Wang, Weixuan
    Liang, Wenxin
    Li, Yuangang
    NEURAL PROCESSING LETTERS, 2020, 51 (01) : 231 - 249
  • [8] Speed Up the Training of Neural Machine Translation
    Xinyue Liu
    Weixuan Wang
    Wenxin Liang
    Yuangang Li
    Neural Processing Letters, 2020, 51 : 231 - 249
  • [9] Minimum Risk Training for Neural Machine Translation
    Shen, Shiqi
    Cheng, Yong
    He, Zhongjun
    He, Wei
    Wu, Hua
    Sun, Maosong
    Liu, Yang
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 1683 - 1692
  • [10] Transformer: A General Framework from Machine Translation to Others
    Yang Zhao
    Jiajun Zhang
    Chengqing Zong
    Machine Intelligence Research, 2023, 20 : 514 - 538