Evaluating Machine Translation Systems with Second Language Proficiency Tests

被引:0
|
作者
Matsuzaki, Takuya [1 ,2 ]
Fujita, Akira [2 ]
Todo, Naoya [2 ]
Arai, Noriko H. [2 ]
机构
[1] Nagoya Univ, Dept Elect Engn & Comp Sci, Nagoya, Aichi, Japan
[2] Natl Inst Informat, Informat & Soc Res Div, Tokyo, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A lightweight, human-in-the-loop evaluation scheme for machine translation (MT) systems is proposed. It extrinsically evaluates MT systems using human subjects' scores on second language ability test problems that are machine-translated to the subjects' native language. A large-scale experiment involving 320 subjects revealed that the context-unawareness of the current MT systems severely damages human performance when solving the test problems, while one of the evaluated MT systems performed as good as a human translation produced in a context-unaware condition. An analysis of the experimental results showed that the extrinsic evaluation captured a different dimension of translation quality than that captured by manual and automatic intrinsic evaluation.
引用
收藏
页码:145 / 149
页数:5
相关论文
共 50 条