Paraphrase Generation with Deep Reinforcement Learning

被引:0
|
作者
Li, Zichao [1 ]
Jiang, Xin [1 ]
Shang, Lifeng [1 ]
Li, Hang [2 ]
机构
[1] Huawei Technol, Noahs Ark Lab, Shenzhen, Peoples R China
[2] Toutiao AI Lab, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic generation of paraphrases from a given sentence is an important yet challenging task in natural language processing (NLP). In this paper, we present a deep reinforcement learning approach to paraphrase generation. Specifically, we propose a new framework for the task, which consists of a generator and an evaluator, both of which are learned from data. The generator, built as a sequence-to-sequence learning model, can produce paraphrases given a sentence. The evaluator, constructed as a deep matching model, can judge whether two sentences are paraphrases of each other. The generator is first trained by deep learning and then further fine-tuned by reinforcement learning in which the reward is given by the evaluator. For the learning of the evaluator, we propose two methods based on supervised learning and inverse reinforcement learning respectively, depending on the type of available training data. Experimental results on two datasets demonstrate the proposed models (the generators) can produce more accurate paraphrases and outperform the state-of-the-art methods in paraphrase generation in both automatic evaluation and human evaluation.
引用
收藏
页码:3865 / 3878
页数:14
相关论文
共 50 条
  • [31] Hierarchical Gait Generation for Modular Robots Using Deep Reinforcement Learning
    Wang, Jiayu
    Hu, Chuxiong
    Zhu, Yu
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS (ICM), 2021,
  • [32] Conformer-RL: A deep reinforcement learning library for conformer generation
    Jiang, Runxuan
    Gogineni, Tarun
    Kammeraad, Joshua
    He, Yifei
    Tewari, Ambuj
    Zimmerman, Paul M.
    [J]. JOURNAL OF COMPUTATIONAL CHEMISTRY, 2022, 43 (27) : 1880 - 1886
  • [33] Animation generation for object transportation with a rope using deep reinforcement learning
    Wong, Sai-Keung
    Wei, Xu-Tao
    [J]. COMPUTER ANIMATION AND VIRTUAL WORLDS, 2023, 34 (3-4)
  • [34] Reinforcement learning-driven deep question generation with rich semantics
    Guan, Menghong
    Mondal, Subrota Kumar
    Dai, Hong-Ning
    Bao, Haiyong
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (02)
  • [35] DEEP REINFORCEMENT LEARNING-BASED AUTOMATIC TEST PATTERN GENERATION
    Li, Wenxing
    Lyu, Hongqin
    Liang, Shengwen
    Liu, Zizhen
    Lin, Ning
    Wang, Zhongrui
    Tian, Pengyu
    Wang, Tiancheng
    Li, Huawei
    [J]. CONFERENCE OF SCIENCE & TECHNOLOGY FOR INTEGRATED CIRCUITS, 2024 CSTIC, 2024,
  • [36] Deep reinforcement learning control of white-light continuum generation
    Valensise, Carlo M.
    Giuseppi, Alessandro
    Cerullo, Giulio
    Polli, Dario
    [J]. OPTICA, 2021, 8 (02): : 239 - 242
  • [37] Automated Gait Generation for Simulated Bodies using Deep Reinforcement Learning
    Ananthakrishnan, Abhishek
    Kanakiya, Vatsal
    Ved, Dipen
    Sharma, Grishma
    [J]. PROCEEDINGS OF THE 2018 SECOND INTERNATIONAL CONFERENCE ON INVENTIVE COMMUNICATION AND COMPUTATIONAL TECHNOLOGIES (ICICCT), 2018, : 90 - 95
  • [38] Automatic Generation Control Based on Deep Reinforcement Learning With Exploration Awareness
    Xi L.
    Yu L.
    Fu Y.
    Huang Y.
    Chen X.
    Kang S.
    [J]. Zhongguo Dianji Gongcheng Xuebao/Proceedings of the Chinese Society of Electrical Engineering, 2019, 39 (14): : 4150 - 4161
  • [39] Contrastive Representation Learning for Exemplar-Guided Paraphrase Generation ☆
    Yang, Haoran
    Lam, Wai
    Li, Piji
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 4754 - 4761
  • [40] Learning to Drive with Deep Reinforcement Learning
    Chukamphaeng, Nut
    Pasupa, Kitsuchart
    Antenreiter, Martin
    Auer, Peter
    [J]. 2021 13TH INTERNATIONAL CONFERENCE ON KNOWLEDGE AND SMART TECHNOLOGY (KST-2021), 2021, : 147 - 152