RESA: Relation Enhanced Self-Attention for Low-Resource Neural Machine Translation

被引:2
|
作者
Wu, Xing [1 ]
Shi, Shumin [1 ]
Huang, Heyan [1 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Low-Resource Neural Machine Translation; Dependency Syntax; Self-Attention;
D O I
10.1109/IALP54817.2021.9675172
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based Neural Machine Translation models have achieved impressive results on many translation tasks. In the meanwhile, some studies prove that extending syntax information can be explicitly incorporated to provide further improvements especially for some low-resource languages. In this paper, we propose RESA: the relation enhanced self-attention for Transformer which can integrate source side dependency syntax. More specifically, dependency parsing produces two kinds of information: dependency heads and relation labels, compared to the previous works only pay attention to dependency heads information, RESA use two methods to integrate relation labels as well: 1) Hard-way that uses a hyper parameter to control the information percentage after mapping relation labels sequence to continuous representations; 2) Gate-way that employs a gate mechanism to mix word information and relation labels information. We evaluate our methods on low-resource Chinese-Tibetan and Chinese-Mongol translation tasks, and the preliminary experimental results show that the proposed model achieves 0.93 and 0.68 BLEU scores gain compared to the baseline model.
引用
收藏
页码:159 / 164
页数:6
相关论文
共 50 条
  • [21] Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures
    Tang, Gongbo
    Mueller, Mathias
    Rios, Annette
    Sennrich, Rico
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4263 - 4272
  • [22] Self-Attention and Dynamic Convolution Hybrid Model for Neural Machine Translation
    Zhang, Zhebin
    Wu, Sai
    Chen, Gang
    Jiang, Dawei
    [J]. 11TH IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG 2020), 2020, : 352 - 359
  • [23] Survey of Low-Resource Machine Translation
    Haddow, Barry
    Bawden, Rachel
    Barone, Antonio Valerio Miceli
    Helcl, Jindrich
    Birch, Alexandra
    [J]. COMPUTATIONAL LINGUISTICS, 2022, 48 (03) : 673 - 732
  • [24] Adapting Attention-Based Neural Network to Low-Resource Mongolian-Chinese Machine Translation
    Wu, Jing
    Hou, Hongxu
    Shen, Zhipeng
    Du, Jian
    Li, Jinting
    [J]. NATURAL LANGUAGE UNDERSTANDING AND INTELLIGENT APPLICATIONS (NLPCC 2016), 2016, 10102 : 470 - 480
  • [25] Acoustic model training using self-attention for low-resource speech recognition
    Park, Hosung
    Kim, Ji-Hwan
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF KOREA, 2020, 39 (05): : 483 - 489
  • [26] Rethinking the Exploitation of Monolingual Data for Low-Resource Neural Machine Translation
    Pang, Jianhui
    Yang, Baosong
    Wong, Derek Fai
    Wan, Yu
    Liu, Dayiheng
    Chao, Lidia Sam
    Xie, Jun
    [J]. COMPUTATIONAL LINGUISTICS, 2023, 50 (01) : 25 - 47
  • [27] Semantic Perception-Oriented Low-Resource Neural Machine Translation
    Wu, Nier
    Hou, Hongxu
    Li, Haoran
    Chang, Xin
    Jia, Xiaoning
    [J]. MACHINE TRANSLATION, CCMT 2021, 2021, 1464 : 51 - 62
  • [28] A Diverse Data Augmentation Strategy for Low-Resource Neural Machine Translation
    Li, Yu
    Li, Xiao
    Yang, Yating
    Dong, Rui
    [J]. INFORMATION, 2020, 11 (05)
  • [29] Neural machine translation for low-resource languages without parallel corpora
    Karakanta, Alina
    Dehdari, Jon
    van Genabith, Josef
    [J]. MACHINE TRANSLATION, 2018, 32 (1-2) : 167 - 189
  • [30] Towards a Low-Resource Neural Machine Translation for Indigenous Languages in Canada
    Ngoc Tan Le
    Sadat, Fatiha
    [J]. TRAITEMENT AUTOMATIQUE DES LANGUES, 2021, 62 (03): : 39 - 63