VERTa: a linguistic approach to automatic machine translation evaluation

被引:7
|
作者
Comelles, Elisabet [1 ]
Atserias, Jordi [2 ]
机构
[1] Univ Barcelona, Gran Via Corts Catalanes 585, E-08007 Barcelona, Spain
[2] Univ Basque Country, Paseo Manuel de Lardizabal 1, Donosti 20018, Spain
关键词
Machine translation; Machine translation evaluation; MT metric; Linguistic features; Qualitative approach;
D O I
10.1007/s10579-018-9430-2
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Machine translation (MT) is directly linked to its evaluation in order to both compare different MT system outputs and analyse system errors so that they can be addressed and corrected. As a consequence, MT evaluation has become increasingly important and popular in the last decade, leading to the development of MT evaluation metrics aiming at automatically assessing MT output. Most of these metrics use reference translations in order to compare system output, and the most well-known and widely spread work at lexical level. In this study we describe and present a linguistically-motivated metric, VERTa, which aims at using and combining a wide variety of linguistic features at lexical, morphological, syntactic and semantic level. Before designing and developing VERTa a qualitative linguistic analysis of data was performed so as to identify the linguistic phenomena that an MT metric must consider (Comelles et al. 2017). In the present study we introduce VERTa's design and architecture and we report the experiments performed in order to develop the metric and to check the suitability and interaction of the linguistic information used. The experiments carried out go beyond traditional correlation scores and step towards a more qualitative approach based on linguistic analysis. Finally, in order to check the validity of the metric, an evaluation has been conducted comparing the metric's performance to that of other well-known state-of-the-art MT metrics.
引用
收藏
页码:57 / 86
页数:30
相关论文
共 50 条
  • [41] An Automatic Evaluation for Online Machine Translation: Holy Quran Case Study
    AlSukhni, Emad
    Al-Kabi, Mohammed N.
    Alsmadi, Izzat M.
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2016, 7 (06) : 118 - 123
  • [42] Automatic Meta-evaluation of Low-Resource Machine Translation Evaluation Metrics
    Yu, Junting
    Liu, Wuying
    He, Hongye
    Wang, Lin
    PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2019, : 136 - 141
  • [43] Effect of Linguistic Information in Neural Machine Translation
    Nakamura, Naomichi
    Isahara, Hitoshi
    2017 4TH INTERNATIONAL CONFERENCE ON ADVANCED INFORMATICS, CONCEPTS, THEORY, AND APPLICATIONS (ICAICTA) PROCEEDINGS, 2017,
  • [44] LOST IN MACHINE TRANSLATION: CONTEXTUAL LINGUISTIC UNCERTAINTY
    Sukhoverkhov, Anton, V
    DeWitt, Dorothy
    Manasidi, Ioannis I.
    Nitta, Keiko
    Krstic, Vladimir
    VESTNIK VOLGOGRADSKOGO GOSUDARSTVENNOGO UNIVERSITETA-SERIYA 2-YAZYKOZNANIE, 2019, 18 (04): : 129 - 144
  • [45] 3 LEVELS OF LINGUISTIC ANALYSIS IN MACHINE TRANSLATION
    ZARECHNAK, M
    JOURNAL OF THE ACM, 1959, 6 (01) : 24 - 32
  • [46] MACHINE TRANSLATION AND LINGUISTIC DATA-PROCESSING
    TOSH, LW
    LINGUISTICS, 1964, (05) : 35 - 39
  • [47] ELU AS A LINGUISTIC DEVELOPMENT ENVIRONMENT FOR MACHINE TRANSLATION
    ESTIVAL, D
    META, 1992, 37 (04) : 693 - 708
  • [48] Machine Translation-Oriented Linguistic Research
    Cai, Lei
    PROCEEDINGS OF THE SECOND NORTHEAST ASIA INTERNATIONAL SYMPOSIUM ON LANGUAGE, LITERATURE AND TRANSLATION, 2012, : 1 - 6
  • [49] Neural Machine Translation as a Novel Approach to Machine Translation
    Benkova, Lucia
    Benko, Lubomir
    DIVAI 2020: 13TH INTERNATIONAL SCIENTIFIC CONFERENCE ON DISTANCE LEARNING IN APPLIED INFORMATICS, 2020, : 499 - 508
  • [50] Automatic machine translation error identification
    de Jesus Martins, Debora Beatriz
    Caseli, Helena de Medeiros
    MACHINE TRANSLATION, 2015, 29 (01) : 1 - 24