Cooperative Self-training of Machine Reading Comprehension

被引:0
|
作者
Luo, Hongyin [1 ]
Li, Shang-Wen [2 ]
Gao, Mingye [3 ]
Yu, Seunghak [2 ]
Glass, James [1 ]
机构
[1] MIT CSAIL, Cambridge, MA 02139 USA
[2] Amazon AI, Bellevue, WA USA
[3] MIT MTL, Cambridge, MA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained language models have significantly improved the performance of downstream language understanding tasks, including extractive question answering, by providing high-quality contextualized word embeddings. However, training question answering models still requires large amounts of annotated data for specific domains. In this work, we propose a cooperative self-training framework, RGX, for automatically generating more non-trivial question-answer pairs to improve model performance. RGX is built upon a masked answer extraction task with an interactive learning environment containing an answer entity Recognizer, a question Generator, and an answer eXtractor. Given a passage with a masked entity, the generator generates a question around the entity, and the extractor is trained to extract the masked entity with the generated question and raw texts. The framework allows the training of question generation and answering models on any text corpora without annotation. We further leverage a self-training technique to improve the performance of both question generation and answer extraction models. Experiment results show that RGX outperforms the state-of-the-art (SOTA) pretrained language models and transfer learning approaches on standard question-answering benchmarks, and yields the new SOTA performance under given model size and transfer learning settings.
引用
收藏
页码:244 / 257
页数:14
相关论文
共 50 条
  • [21] KUDOS FOR SELF-TRAINING AIDS
    BRYANT, SF
    COMPUTER DECISIONS, 1984, 16 (14): : 44 - &
  • [22] Doubly Robust Self-Training
    Zhu, Banghua
    Ding, Mingyu
    Jacobson, Philip
    Wu, Ming
    Zhan, Wei
    Jordan, Michael I.
    Jiao, Jiantao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [23] Deep Bayesian Self-Training
    Ribeiro, Fabio De Sousa
    Caliva, Francesco
    Swainson, Mark
    Gudmundsson, Kjartan
    Leontidis, Georgios
    Kollias, Stefanos
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (09): : 4275 - 4291
  • [24] RECURSIVE SELF-TRAINING ALGORITHMS
    TSYPKIN, YZ
    KELMANS, GK
    ENGINEERING CYBERNETICS, 1967, (05): : 70 - &
  • [25] Rethinking Pre-training and Self-training
    Zoph, Barret
    Ghiasi, Golnaz
    Lin, Tsung-Yi
    Cui, Yin
    Liu, Hanxiao
    Cubuk, Ekin D.
    Le, Quoc V.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [26] NICT Self-Training Approach to Neural Machine Translation at NMT-2018
    Imamura, Kenji
    Sumita, Eiichiro
    NEURAL MACHINE TRANSLATION AND GENERATION, 2018, : 110 - 115
  • [27] Research on self-training neural machine translation based on monolingual priority sampling
    Zhang X.
    Pang L.
    Du X.
    Lu T.
    Xia Y.
    Tongxin Xuebao/Journal on Communications, 2024, 45 (04): : 65 - 72
  • [28] Improving Machine Reading Comprehension through A Simple Masked-Training Scheme
    Yao, Xun
    Ma, Junlong
    Hu, Xinrong
    Yang, Jie
    Li, Yuan-Fang
    13TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING AND THE 3RD CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, IJCNLP-AACL 2023, 2023, : 222 - 232
  • [29] Combining permuted language model and adversarial training for Chinese machine reading comprehension
    Liu J.
    Chu X.
    Wang J.
    Wang M.
    Wang Y.
    Journal of Intelligent and Fuzzy Systems, 2024, 46 (04): : 10059 - 10073
  • [30] Self-Training System of Calligraphy Brushwork
    Morikawa, Ami
    Tsuda, Naoaki
    Nomura, Yoshihiko
    Kato, Norihiko
    COMPANION OF THE 2017 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI'17), 2017, : 215 - 216