VERTa: a linguistic approach to automatic machine translation evaluation

被引:7
|
作者
Comelles, Elisabet [1 ]
Atserias, Jordi [2 ]
机构
[1] Univ Barcelona, Gran Via Corts Catalanes 585, E-08007 Barcelona, Spain
[2] Univ Basque Country, Paseo Manuel de Lardizabal 1, Donosti 20018, Spain
关键词
Machine translation; Machine translation evaluation; MT metric; Linguistic features; Qualitative approach;
D O I
10.1007/s10579-018-9430-2
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Machine translation (MT) is directly linked to its evaluation in order to both compare different MT system outputs and analyse system errors so that they can be addressed and corrected. As a consequence, MT evaluation has become increasingly important and popular in the last decade, leading to the development of MT evaluation metrics aiming at automatically assessing MT output. Most of these metrics use reference translations in order to compare system output, and the most well-known and widely spread work at lexical level. In this study we describe and present a linguistically-motivated metric, VERTa, which aims at using and combining a wide variety of linguistic features at lexical, morphological, syntactic and semantic level. Before designing and developing VERTa a qualitative linguistic analysis of data was performed so as to identify the linguistic phenomena that an MT metric must consider (Comelles et al. 2017). In the present study we introduce VERTa's design and architecture and we report the experiments performed in order to develop the metric and to check the suitability and interaction of the linguistic information used. The experiments carried out go beyond traditional correlation scores and step towards a more qualitative approach based on linguistic analysis. Finally, in order to check the validity of the metric, an evaluation has been conducted comparing the metric's performance to that of other well-known state-of-the-art MT metrics.
引用
收藏
页码:57 / 86
页数:30
相关论文
共 50 条
  • [31] Linguistic evaluation of German-English Machine Translation using a Test Suite
    Avramidis, Eleftherios
    Macketanz, Vivien
    Strohriegel, Ursula
    Uszkoreit, Hans
    FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), 2019, : 445 - 454
  • [32] A New Evaluation Approach for Sign Language Machine Translation
    Almohimeed, Abdulaziz
    Wald, Mike
    Damper, Robert
    ASSISTIVE TECHNOLOGY FROM ADAPTED EQUIPMENT TO INCLUSIVE ENVIRONMENTS, 2009, 25 : 498 - 502
  • [33] Applying the Cognitive Machine Translation Evaluation Approach to Arabic
    Temnikova, Irina
    Zaghouani, Wajdi
    Vogel, Stephan
    Habash, Nizar
    LREC 2016 - TENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2016, : 3644 - 3651
  • [34] Detecting errors in machine translation using residuals and metrics of automatic evaluation
    Munk, Michal
    Munkova, Dasa
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 34 (05) : 3211 - 3223
  • [35] STD: An Automatic Evaluation Metric for Machine Translation Based on Word Embeddings
    Li, Pairui
    Chen, Chuan
    Zheng, Wujie
    Deng, Yuetang
    Ye, Fanghua
    Zheng, Zibin
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (10) : 1497 - 1506
  • [36] A Comprehensive Survey on Various Fully Automatic Machine Translation Evaluation Metrics
    Chauhan, Shweta
    Daniel, Philemon
    NEURAL PROCESSING LETTERS, 2023, 55 (09) : 12663 - 12717
  • [37] BLONDE: An Automatic Evaluation Metric for Document-level Machine Translation
    Jiang, Yuchen Eleanor
    Liu, Tianyu
    Ma, Shuming
    Zhang, Dongdong
    Yang, Jian
    Huang, Haoyang
    Sennrich, Rico
    Sachan, Mrinmaya
    Cotterell, Ryan
    Zhou, Ming
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1550 - 1565
  • [38] Machine Translation Evaluation: Manual Versus Automatic-A Comparative Study
    Maurya, Kaushal Kumar
    Ravindran, Renjith P.
    Anirudh, Ch Ram
    Murthy, Kavi Narayana
    DATA ENGINEERING AND COMMUNICATION TECHNOLOGY, ICDECT-2K19, 2020, 1079 : 541 - 553
  • [39] A Comprehensive Survey on Various Fully Automatic Machine Translation Evaluation Metrics
    Shweta Chauhan
    Philemon Daniel
    Neural Processing Letters, 2023, 55 : 12663 - 12717
  • [40] Filtering Pseudo-References by Paraphrasing for Automatic Evaluation of Machine Translation
    Yoshimura, Ryoma
    Shimanaka, Hiroki
    Matsumura, Yukio
    Yamagishi, Hayahide
    Komachi, Mamoru
    FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), 2019, : 521 - 525