Document-Level Machine Translation with Large Language Models

被引:0
|
作者
Wang, Longyue [1 ]
Lyu, Chenyang [2 ]
Ji, Tianbo [3 ]
Zhang, Zhirui [1 ]
Yu, Dian [1 ]
Shi, Shuming [1 ]
Tu, Zhaopeng [1 ]
机构
[1] Tencent AI Lab, Shenzhen, Guangdong, Peoples R China
[2] MBZUAI, Abu Dhabi, U Arab Emirates
[3] Dublin City Univ, Dublin, Ireland
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) such as ChatGPT can produce coherent, cohesive, relevant, and fluent answers for various natural language processing (NLP) tasks. Taking documentlevel machine translation (MT) as a testbed, this paper provides an in-depth evaluation of LLMs' ability on discourse modeling. The study focuses on three aspects: 1) Effects of Context-Aware Prompts, where we investigate the impact of different prompts on document-level translation quality and discourse phenomena; 2) Comparison of Translation Models, where we compare the translation performance of ChatGPT with commercial MT systems and advanced document-level MT methods; 3) Analysis of Discourse Modelling Abilities, where we further probe discourse knowledge encoded in LLMs and shed light on impacts of training techniques on discourse modeling. By evaluating on a number of benchmarks, we surprisingly find that LLMs have demonstrated superior performance and show potential to become a new paradigm for document-level translation: 1) leveraging their powerful long-text modeling capabilities, GPT-3.5 and GPT-4 outperform commercial MT systems in terms of human evaluation;(1) 2) GPT-4 demonstrates a stronger ability for probing linguistic knowledge than GPT-3.5. This work highlights the challenges and opportunities of LLMs for MT, which we hope can inspire the future design and evaluation of LLMs.(2)
引用
收藏
页码:16646 / 16661
页数:16
相关论文
共 50 条
  • [21] Document-Level Neural Machine Translation with Hierarchical Attention Networks
    Miculicich, Lesly
    Dhananjay, Ram
    Pappas, Nikolaos
    Henderson, James
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2947 - 2954
  • [22] A Survey on Document-level Neural Machine Translation: Methods and Evaluation
    Maruf, Sameen
    Saleh, Fahimeh
    Haffari, Gholamreza
    ACM COMPUTING SURVEYS, 2021, 54 (02)
  • [23] Document-Level Neural Machine Translation with Associated Memory Network
    Jiang, Shu
    Wang, Rui
    Li, Zuchao
    Utiyama, Masao
    Chen, Kehai
    Sumita, Eiichiro
    Zhao, Hai
    Lu, Bao-liang
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (10) : 1712 - 1723
  • [24] Modeling Discourse Structure for Document-level Neural Machine Translation
    Chen, Junxuan
    Li, Xiang
    Zhang, Jiarui
    Zhou, Chulun
    Cui, Jianwei
    Wang, Bin
    Su, Jinsong
    WORKSHOP ON AUTOMATIC SIMULTANEOUS TRANSLATION CHALLENGES, RECENT ADVANCES, AND FUTURE DIRECTIONS, 2020, : 30 - 36
  • [25] Improving Document-Level Neural Machine Translation with Domain Adaptation
    Ul Haq, Sami
    Rauf, Sadaf Abdul
    Shoukat, Arslan
    Noor-e-Hira
    NEURAL GENERATION AND TRANSLATION, 2020, : 225 - 231
  • [26] CONTEXT-ADAPTIVE DOCUMENT-LEVEL NEURAL MACHINE TRANSLATION
    Zhang, Linlin
    Zhang, Zhirui
    Chen, Boxing
    Luo, Weihua
    Si, Luo
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6232 - 6236
  • [27] Target-Side Augmentation for Document-Level Machine Translation
    Bao, Guangsheng
    Teng, Zhiyang
    Zhang, Yue
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 10725 - 10740
  • [28] A Simple and Effective Unified Encoder for Document-Level Machine Translation
    Ma, Shuming
    Zhang, Dongdong
    Zhou, Ming
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3505 - 3511
  • [29] BLONDE: An Automatic Evaluation Metric for Document-level Machine Translation
    Jiang, Yuchen Eleanor
    Liu, Tianyu
    Ma, Shuming
    Zhang, Dongdong
    Yang, Jian
    Huang, Haoyang
    Sennrich, Rico
    Sachan, Mrinmaya
    Cotterell, Ryan
    Zhou, Ming
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1550 - 1565
  • [30] Document Flattening: Beyond Concatenating Context for Document-Level Neural Machine Translation
    Wu, Minghao
    Foster, George
    Qu, Lizhen
    Haffari, Gholamreza
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 448 - 462