Self-Regulated Interactive Sequence-to-Sequence Learning

被引:0
|
作者
Kreutzer, Julia [1 ]
Riezler, Stefan [1 ,2 ]
机构
[1] Heidelberg Univ, Computat Linguist, Heidelberg, Germany
[2] Heidelberg Univ, IWR, Heidelberg, Germany
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Not all types of supervision signals are created equal: Different types of feedback have different costs and effects on learning. We show how self-regulation strategies that decide when to ask for which kind of feedback from a teacher (or from oneself) can be cast as a learning-to-learn problem leading to improved cost-aware sequence-to-sequence learning. In experiments on interactive neural machine translation, we find that the self-regulator discovers an epsilon-greedy strategy for the optimal cost-quality trade-off by mixing different feedback types including corrections, error markups, and self-supervision. Furthermore, we demonstrate its robustness under domain shift and identify it as a promising alternative to active learning.
引用
收藏
页码:303 / 315
页数:13
相关论文
共 50 条
  • [1] Sequence-to-sequence self calibration
    Wolf, L
    Zomet, A
    [J]. COMPUTER VISION - ECCV 2002, PT II, 2002, 2351 : 370 - 382
  • [2] Semantic Matching for Sequence-to-Sequence Learning
    Zhang, Ruiyi
    Chen, Changyou
    Zhang, Xinyuan
    Bai, Ke
    Carin, Lawrence
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 212 - 222
  • [3] Sequence-to-sequence modeling for graph representation learning
    Taheri, Aynaz
    Gimpel, Kevin
    Berger-Wolf, Tanya
    [J]. APPLIED NETWORK SCIENCE, 2019, 4 (01)
  • [4] Incorporating Copying Mechanism in Sequence-to-Sequence Learning
    Gu, Jiatao
    Lu, Zhengdong
    Li, Hang
    Li, Victor O. K.
    [J]. PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 1631 - 1640
  • [5] Token-Level Self-Evolution Training for Sequence-to-Sequence Learning
    Peng, Keqin
    Ding, Liang
    Zhong, Qihuang
    Ouyang, Yuanxin
    Rong, Wenge
    Xiong, Zhang
    Tao, Dacheng
    [J]. 61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 841 - 850
  • [6] Matrix product operators for sequence-to-sequence learning
    Guo, Chu
    Jie, Zhanming
    Lu, Wei
    Poletti, Dario
    [J]. PHYSICAL REVIEW E, 2018, 98 (04)
  • [7] Deep Reinforcement Learning for Sequence-to-Sequence Models
    Keneshloo, Yaser
    Shi, Tian
    Ramakrishnan, Naren
    Reddy, Chandan K.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (07) : 2469 - 2489
  • [8] Sequence-to-Sequence Learning with Latent Neural Grammars
    Kim, Yoon
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [9] DUALFORMER: A UNIFIED BIDIRECTIONAL SEQUENCE-TO-SEQUENCE LEARNING
    Chien, Jen-Tzung
    Chang, Wei-Hsiang
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7718 - 7722
  • [10] Learning Damage Representations with Sequence-to-Sequence Models
    Yang, Qun
    Shen, Dejian
    [J]. SENSORS, 2022, 22 (02)