Domain-Aware Self-Attention for Multi-Domain Neural Machine Translation

被引:2
|
作者
Zhang, Shiqi [1 ]
Liu, Yan [2 ]
Xiong, Deyi [2 ]
Zhang, Pei [1 ]
Chen, Boxing [1 ]
机构
[1] Alibaba Grp Inc, Hong Kong, Peoples R China
[2] Tianjin Univ, Coll Intelligence & Comp, Tianjin, Peoples R China
来源
关键词
Transformer; machine translation; domain adaptation; unsupervised learning;
D O I
10.21437/Interspeech.2021-1477
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
In this paper, we investigate multi-domain neural machine translation (NMT) that translates sentences of different domains in a single model. To this end, we propose a domain-aware self-attention mechanism that jointly learns domain representations with the single NMT model. The learned domain representations are integrated into both the encoder and decoder. We further propose two different domain representation learning approaches: 1) word-level unsupervised learning via a domain attention network and 2) guided learning with an auxiliary loss. The two learning approaches allow our multi-domain NMT to work in different settings as to whether the domain information is available or not. Experiments on both Chinese-English and English-French demonstrate that our multi-domain model outperforms a strong baseline built on the Transformer and other previous multi-domain NMT approaches. Further analyses show that our model is able to learn domain clusters even without prior knowledge about the domain structure.
引用
收藏
页码:2047 / 2051
页数:5
相关论文
共 50 条
  • [1] Multi-Domain Sentiment Classification Based on Domain-Aware Embedding and Attention
    Cai, Yitao
    Wan, Xiaojun
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4904 - 4910
  • [2] Self-attention StarGAN for Multi-domain Image-to-Image Translation
    He, Ziliang
    Yang, Zhenguo
    Mao, Xudong
    Lv, Jianming
    Li, Qing
    Liu, Wenyin
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: IMAGE PROCESSING, PT III, 2019, 11729 : 537 - 549
  • [3] Unsupervised Domain Adaptation for Neural Machine Translation with Domain-Aware Feature Embeddings
    Dou, Zi-Yi
    Hu, Junjie
    Anastasopoulos, Antonios
    Neubig, Graham
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1417 - 1422
  • [4] Domain-Aware Dialogue State Tracker for Multi-Domain Dialogue Systems
    Balaraman, Vevake
    Magnini, Bernardo
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 866 - 873
  • [5] Domain-Aware Contrastive Knowledge Transfer for Multi-domain Imbalanced Data
    Ke, Zixuan
    Kachuee, Mohammad
    Lee, Sungjin
    [J]. PROCEEDINGS OF THE 12TH WORKSHOP ON COMPUTATIONAL APPROACHES TO SUBJECTIVITY, SENTIMENT & SOCIAL MEDIA ANALYSIS, 2022, : 25 - 36
  • [6] Domain Adaptation and Multi-Domain Adaptation for Neural Machine Translation: A Survey
    Saunders, Danielle
    [J]. Journal of Artificial Intelligence Research, 2022, 75 : 351 - 424
  • [7] Domain Adaptation and Multi-Domain Adaptation for Neural Machine Translation: A Survey
    Saunders, Danielle
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2022, 75 : 351 - 424
  • [8] Factorized Transformer for Multi-Domain Neural Machine Translation
    Deng, Yongchao
    Yu, Hongfei
    Yu, Heng
    Duan, Xiangyu
    Luo, Weihua
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4221 - 4230
  • [9] Uncertainty-Aware Balancing for Multilingual and Multi-Domain Neural Machine Translation Training
    Wu, Minghao
    Li, Yitong
    Zhang, Meng
    Li, Liangyou
    Haffari, Gholamreza
    Liu, Qun
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 7291 - 7305
  • [10] Multi-Granularity Self-Attention for Neural Machine Translation
    Hao, Jie
    Wang, Xing
    Shi, Shuming
    Zhang, Jinfeng
    Tu, Zhaopeng
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 887 - 897