Document-Level Neural Machine Translation With Recurrent Context States

被引:1
|
作者
Zhao, Yue [1 ]
Liu, Hui [2 ]
机构
[1] Northeastern Univ, Sch Marxism, Shenyang 110819, Peoples R China
[2] Northeastern Univ, Sch Comp Sci & Engn, Shenyang 110819, Peoples R China
关键词
Context modeling; Training; Complexity theory; Decoding; Computational modeling; Machine translation; Transformers; Neural machine translation; document-level translation; speeding up;
D O I
10.1109/ACCESS.2023.3247508
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Integrating contextual information into sentence-level neural machine translation (NMT) systems has been proven to be effective in generating fluent and coherent translations. However, taking too much context into account slows down these systems, especially when context-aware models are applied to the decoder side. To improve efficiency, we propose a simple and fast method to encode all sentences in an arbitrary large context window. It makes contextual representations in the process of translating each sentence so that the overhead introduced by the context model is almost negligible. We experiment with our method on three widely used English-German document-level translation datasets, which obtain substantial improvements over the sentence-level baseline with almost no loss in efficiency. Moreover, our method also achieves comparable performance with previous strong context-aware baselines and speeds up the inference by 1.53x. The speed-up is even larger when more contexts are taken into account. On the ContraPro pronoun translation dataset, it significantly outperforms the strong baseline.
引用
收藏
页码:27519 / 27526
页数:8
相关论文
共 50 条
  • [1] CONTEXT-ADAPTIVE DOCUMENT-LEVEL NEURAL MACHINE TRANSLATION
    Zhang, Linlin
    Zhang, Zhirui
    Chen, Boxing
    Luo, Weihua
    Si, Luo
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6232 - 6236
  • [2] Document Flattening: Beyond Concatenating Context for Document-Level Neural Machine Translation
    Wu, Minghao
    Foster, George
    Qu, Lizhen
    Haffari, Gholamreza
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 448 - 462
  • [3] Routing Based Context Selection for Document-Level Neural Machine Translation
    Fei, Weilun
    Jian, Ping
    Zhu, Xiaoguang
    Lin, Yi
    MACHINE TRANSLATION, CCMT 2021, 2021, 1464 : 77 - 91
  • [4] Document-Level Neural Machine Translation with Hierarchical Modeling of Global Context
    Tan, Xin
    Zhang, Long-Yin
    Zhou, Guo-Dong
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2022, 37 (02) : 295 - 308
  • [5] Toward Understanding Most of the Context in Document-Level Neural Machine Translation
    Choi, Gyu-Hyeon
    Shin, Jong-Hun
    Lee, Yo-Han
    Kim, Young-Kil
    ELECTRONICS, 2022, 11 (15)
  • [6] Document-level Neural Machine Translation Using BERT as Context Encoder
    Guo, Zhiyu
    Minh Le Nguyen
    AACL-IJCNLP 2020: THE 1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2020, : 94 - 100
  • [7] Document-Level Neural Machine Translation with Hierarchical Modeling of Global Context
    Xin Tan
    Long-Yin Zhang
    Guo-Dong Zhou
    Journal of Computer Science and Technology, 2022, 37 : 295 - 308
  • [8] Hierarchical Modeling of Global Context for Document-Level Neural Machine Translation
    Tan, Xin
    Zhang, Longyin
    Xiong, Deyi
    Zhou, Guodong
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1576 - 1585
  • [9] Rethinking Document-level Neural Machine Translation
    Sun, Zewei
    Wang, Mingxuan
    Zhou, Hao
    Zhao, Chengqi
    Huang, Shujian
    Chen, Jiajun
    Li, Lei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3537 - 3548
  • [10] Document-Level Adaptation for Neural Machine Translation
    Kothur, Sachith Sri Ram
    Knowles, Rebecca
    Koehn, Philipp
    NEURAL MACHINE TRANSLATION AND GENERATION, 2018, : 64 - 73