Document-Level Machine Translation with Large Language Models

被引:0
|
作者
Wang, Longyue [1 ]
Lyu, Chenyang [2 ]
Ji, Tianbo [3 ]
Zhang, Zhirui [1 ]
Yu, Dian [1 ]
Shi, Shuming [1 ]
Tu, Zhaopeng [1 ]
机构
[1] Tencent AI Lab, Shenzhen, Guangdong, Peoples R China
[2] MBZUAI, Abu Dhabi, U Arab Emirates
[3] Dublin City Univ, Dublin, Ireland
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) such as ChatGPT can produce coherent, cohesive, relevant, and fluent answers for various natural language processing (NLP) tasks. Taking documentlevel machine translation (MT) as a testbed, this paper provides an in-depth evaluation of LLMs' ability on discourse modeling. The study focuses on three aspects: 1) Effects of Context-Aware Prompts, where we investigate the impact of different prompts on document-level translation quality and discourse phenomena; 2) Comparison of Translation Models, where we compare the translation performance of ChatGPT with commercial MT systems and advanced document-level MT methods; 3) Analysis of Discourse Modelling Abilities, where we further probe discourse knowledge encoded in LLMs and shed light on impacts of training techniques on discourse modeling. By evaluating on a number of benchmarks, we surprisingly find that LLMs have demonstrated superior performance and show potential to become a new paradigm for document-level translation: 1) leveraging their powerful long-text modeling capabilities, GPT-3.5 and GPT-4 outperform commercial MT systems in terms of human evaluation;(1) 2) GPT-4 demonstrates a stronger ability for probing linguistic knowledge than GPT-3.5. This work highlights the challenges and opportunities of LLMs for MT, which we hope can inspire the future design and evaluation of LLMs.(2)
引用
收藏
页码:16646 / 16661
页数:16
相关论文
共 50 条
  • [41] Semantically Constrained Document-Level Chinese-Mongolian Neural Machine Translation
    Li, Haoran
    Hou, Hongxu
    Wu, Nier
    Jia, Xiaoning
    Chang, Xin
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [42] Has Machine Translation Achieved Human Parity? A Case for Document-level Evaluation
    Laeubli, Samuel
    Sennrich, Rico
    Volk, Martin
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4791 - 4796
  • [43] A Document-Level Machine Translation Quality Estimation Model Based on Centering Theory
    Chen, Yidong
    Zhong, Enjun
    Tong, Yiqi
    Qiu, Yanru
    Shi, Xiaodong
    MACHINE TRANSLATION, CCMT 2021, 2021, 1464 : 1 - 15
  • [44] Importance-Aware Data Augmentation for Document-Level Neural Machine Translation
    Wu, Minghao
    Wang, Yufei
    Foster, George
    Qiu, Lizhen
    Haffari, Gholamreza
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 740 - 752
  • [45] Document-Level Machine Translation Evaluation Metrics Enhanced with Simplified Lexical Chain
    Gong, Zhengxian
    Zhou, Guodong
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2015, 2015, 9362 : 396 - 403
  • [46] Investigating Contextual Influence in Document-Level Translation
    Nayak, Prashanth
    Haque, Rejwanul
    Kelleher, John D.
    Way, Andy
    INFORMATION, 2022, 13 (05)
  • [47] Context-aware Decoder for Neural Machine Translation using a Target-side Document-Level Language Model
    Sugiyama, Amane
    Yoshinaga, Naoki
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5781 - 5791
  • [48] Document-Level Machine Translation - Ensuring Translational Consistency of Non-Local Phenomena
    Martinez Garcia, Eva
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2021, (66): : 197 - 200
  • [49] Dynamic Context Selection for Document-level Neural Machine Translation via Reinforcement Learning
    Kang, Xiaomian
    Zhao, Yang
    Zhang, Jiajun
    Zong, Chengqing
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2242 - 2254
  • [50] Injecting User Identity Into Pretrained Language Models for Document-Level Sentiment Classification
    Cao, Xinlei
    Yu, Jinyang
    Zhuang, Yan
    IEEE ACCESS, 2022, 10 : 30157 - 30167