Syntax-aware Transformer Encoder for Neural Machine Translation

被引:0
|
作者
Duan, Sufeng [1 ]
Zhao, Hai [1 ]
Zhou, Junru [1 ]
Wang, Rui [2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Key Lab Shanghai Educ Commiss Intelligent Interac, MoE Key Lab Artificial Intelligence,AI Inst, Shanghai, Peoples R China
[2] Natl Inst Informat & Commun Technol NICT, Kyoto, Japan
基金
中国国家自然科学基金;
关键词
Neural Machine Translation; dependency parsing; POS Tagging;
D O I
10.1109/ialp48816.2019.9037672
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Syntax has been shown a helpful clue in various natural language processing tasks including previous statistical machine translation and recurrent neural network based machine translation. However, since the state-of-the-art neural machine translation (NMT) has to be built on the Transformer based encoder, few attempts are found on such a syntax enhancement. Thus in this paper, we explore effective ways to introduce syntax into Transformer for better machine translation. We empirically compare two ways, positional encoding and input embedding, to exploit syntactic clues from dependency tree over source sentence. Our proposed methods have a merit keeping the architecture of Transformer unchanged, thus the efficiency of Transformer can be kept. The experimental results on IWSLT' 14 German-to-English and WMT14 English-to-German show that our method can yield advanced results over strong Transformer baselines.
引用
收藏
页码:396 / 401
页数:6
相关论文
共 50 条
  • [31] Syntax-aware Semantic Role Labeling without Parsing
    Cai R.
    Lapata M.
    Transactions of the Association for Computational Linguistics, 2019, 7 : 343 - 356
  • [32] Context, Structure and Syntax-aware RST Discourse Parsing
    Desai, Takshak
    Moldovan, Dan, I
    2021 IEEE 15TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC 2021), 2021, : 155 - 162
  • [33] A Convolutional Encoder Model for Neural Machine Translation
    Gehring, Jonas
    Auli, Michael
    Grangier, David
    Dauphin, Yann N.
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 123 - 135
  • [34] Improved Neural Machine Translation with Source Syntax
    Wu, Shuangzhi
    Zhou, Ming
    Zhang, Dongdong
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4179 - 4185
  • [35] Modeling Source Syntax for Neural Machine Translation
    Li, Junhui
    Xiong, Deyi
    Tu, Zhaopeng
    Zhu, Muhua
    Zhang, Min
    Zhou, Guodong
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 688 - 697
  • [36] Improve Neural Machine Translation by Syntax Tree
    Chen, Siyu
    Yu, Qingsong
    ISCSIC'18: PROCEEDINGS OF THE 2ND INTERNATIONAL SYMPOSIUM ON COMPUTER SCIENCE AND INTELLIGENT CONTROL, 2018,
  • [37] SATD: syntax-aware handwritten mathematical expression recognition based on tree-structured transformer decoder
    Fu, Pengbin
    Xiao, Ganyun
    Yang, Huirong
    VISUAL COMPUTER, 2024, 41 (2): : 883 - 900
  • [38] FastGraphTTS: An Ultrafast Syntax-Aware Speech Synthesis Framework
    Wang, Jianzong
    Zhang, Xulong
    Sun, Aolan
    Cheng, Ning
    Xiao, Jing
    2023 IEEE 35TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2023, : 905 - 912
  • [39] Dependency-based syntax-aware word representations
    Zhang, Meishan
    Li, Zhenghua
    Fu, Guohong
    Zhang, Min
    ARTIFICIAL INTELLIGENCE, 2021, 292
  • [40] Syntax-Aware Network for Handwritten Mathematical Expression Recognition
    Yuan, Ye
    Liu, Xiao
    Dikubab, Wondimu
    Liu, Hui
    Ji, Zhilong
    Wu, Zhongqin
    Bai, Xiang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 4543 - 4552