Transformer-based Automatic Post-Editing Model with Joint Encoder and Multi-source Attention of Decoder

被引:0
|
作者
Lee, WonKee [1 ]
Shin, Jaehun [1 ]
Lee, Jong-Hyeok [1 ]
机构
[1] Pohang Univ Sci & Technol POSTECH, Dept Comp Sci & Engn, Pohang, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes POSIECH's submission to the WMT 2019 shared task on Automatic Post-Editing (APE). In this paper, we propose a new multi-source APE model by extending Transformer. The main contributions of our study are that we 1) reconstruct the encoder to generate a joint representation of translation (mt) and its src context, in addition to the conventional src encoding and 2) suggest two types of multi-source attention layers to compute attention between two outputs of the encoder and the decoder state in the decoder. Furthermore, we train our model by applying various teacher-forcing ratios to alleviate exposure bias. Finally, we adopt the ensemble technique across variations of our model. Experiments on the WMT19 English-German APE data set show improvements in terms of both TER and BLEU scores over the baseline. Our primary submission achieves -0.73 in TER and +1.49 in BLEU compared to the baseline, and ranks second among all submitted systems.
引用
收藏
页码:112 / 117
页数:6
相关论文
共 41 条
  • [1] Exploration of Effective Attention Strategies for Neural Automatic Post-editing with Transformer
    Shin, Jaehun
    Lee, Wonkee
    Go, Byung-Hyun
    Jung, Baikjin
    Kim, Youngkil
    Lee, Jong-Hyeok
    [J]. ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (06)
  • [2] Transformer-based Encoder-Decoder Model for Surface Defect Detection
    Lu, Xiaofeng
    Fan, Wentao
    [J]. 6TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE, ICIAI2022, 2022, : 125 - 130
  • [3] Transformer-Based Multi-Source Domain Adaptation Without Source Data
    Li, Gang
    Wu, Chao
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [4] A Hybrid Machine Transliteration Model Based on Multi-source Encoder–Decoder Framework: English to Manipuri
    Lenin Laitonjam
    Sanasam Ranbir Singh
    [J]. SN Computer Science, 2022, 3 (2)
  • [5] Unbabel's Submission to the WMT2019 APE Shared Task: BERT-based Encoder-Decoder for Automatic Post-Editing
    Lopes, Antonio, V
    Farajian, M. Amin
    Correia, Goncalo M.
    Trenous, Jonay
    Martins, Andre F. T.
    [J]. FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), VOL 3: SHARED TASK PAPERS, DAY 2, 2019, : 118 - 123
  • [6] Seq2Code: Transformer-Based Encoder-Decoder Model for Python']Python Source Code Generation
    Laskari, Naveen Kumar
    Reddy, K. Adi Narayana
    Reddy, M. Indrasena
    [J]. THIRD CONGRESS ON INTELLIGENT SYSTEMS, CIS 2022, VOL 1, 2023, 608 : 301 - 309
  • [7] Diverter transformer-based multi-encoder-multi-decoder network model for medical retinal blood vessel image segmentation
    Wu, Chengwei
    Guo, Min
    Ma, Miao
    Wang, Kaiguang
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 93
  • [8] Automatic Generation of Chinese Couplets with Attention Based Encoder-Decoder Model
    Yuan, Shengqiong
    Zhong, Luo
    Li, Lin
    Zhang, Rui
    [J]. 2019 2ND IEEE CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2019), 2019, : 65 - 70
  • [9] A Transformer-based Multi-modal Joint Attention Fusion Model for Molecular Property Prediction
    Wang, Ke
    Zhang, Wei
    Liu, Yong
    [J]. Proceedings - 2023 2023 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2023, 2023, : 4972 - 4974
  • [10] Transformer-based multi-source images instance segmentation network for composite materials
    Ke Y.
    Fu Y.
    Zhou W.
    Zhu W.
    [J]. Hongwai yu Jiguang Gongcheng/Infrared and Laser Engineering, 2023, 52 (02):