Improve example-based machine translation quality for low-resource language using ontology

被引:1
|
作者
Khan Md Anwarus K.M.A. [1 ]
Yamada S. [2 ]
Tetsuro N. [3 ]
机构
[1] IBM Research Tokyo, 19-21 Nihonbashi, Hakozaki-cho, Chuo-ku, Tokyo
[2] NTT Corporation, NTT Hibiya Building, 1-1-6 Uchisaiwai-cho, Chiyoda-ku, Tokyo
[3] University of Electro-Communications, Graduate School of Informatics and Engineering, 1-5-1 Chofugaoka, Chofu, Tokyo
关键词
Example-based machine translation; Knowledge engineering; WordNet;
D O I
10.2991/ijndc.2017.5.3.6
中图分类号
学科分类号
摘要
In this research we propose to use ontology to improve the performance of an EBMT system for low-resource language pair. The EBMT architecture use chunk-string templates (CSTs) and unknown word translation mechanism. CSTs consist of a chunk in source-language, a string in target-language, and word alignment in-formation. For unknown word translation, we used WordNet hypernym tree and English-Bengali dictionary. CSTs improved the wide-coverage by 57 points and quality by 48.81 points in human evaluation. Currently 64.29% of the test-set translations by the system were acceptable. The combined solutions of CSTs and unknown words generated 67.85% acceptable translations from the test-set. Un-known words mechanism improved translation quality by 3.56 points in human evaluation. Copyright © 2017, the Authors.
引用
收藏
页码:176 / 191
页数:15
相关论文
共 50 条
  • [21] A Survey on Low-Resource Neural Machine Translation
    Wang, Rui
    Tan, Xu
    Luo, Renqian
    Qin, Tao
    Liu, Tie-Yan
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4636 - 4643
  • [22] A Survey on Low-resource Neural Machine Translation
    Li H.-Z.
    Feng C.
    Huang H.-Y.
    Huang, He-Yan (hhy63@bit.edu.cn), 1600, Science Press (47): : 1217 - 1231
  • [23] Transformers for Low-resource Neural Machine Translation
    Gezmu, Andargachew Mekonnen
    Nuernberger, Andreas
    ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 1, 2022, : 459 - 466
  • [24] The Low-Resource Double Bind: An Empirical Study of Pruning for Low-Resource Machine Translation
    Ahia, Orevaoghene
    Kreutzer, Julia
    Hooker, Sara
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3316 - 3333
  • [25] Merging example-based and statistical machine translation: An experiment
    Langlais, P
    Simard, M
    MACHINE TRANSLATION: FROM RESEARCH TO REAL USERS, 2002, 2499 : 104 - 113
  • [26] Recent research in the field of example-based machine translation
    Carl, M
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, 2001, 2004 : 195 - 196
  • [27] Computing Transfer Score in Example-Based Machine Translation
    Jaworski, Rafal
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, 2010, 6008 : 406 - 416
  • [28] GATITOS: Using a New Multilingual Lexicon for Low-resource Machine Translation
    Jones, Alex
    Caswell, Isaac
    Saxena, Ishank
    Firat, Orhan
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 371 - 405
  • [29] Combining invertible Example-Based Machine Translation with translation memory technology
    Carl, M
    ENVISIONING MACHINE TRANSLATION IN THE INFORMATION FUTURE, PROCEEDINGS, 2000, 1934 : 127 - 136
  • [30] Generation of Translation Tables Adequate for Example-Based Machine Translation by Analogy
    Kimura, Tatsuya
    Matsuoka, Jin
    Nishikawa, Yusuke
    Lepage, Yves
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND SOFTWARE ENGINEERING (AISE 2014), 2014, : 200 - 203