Paraphrase Generation with Deep Reinforcement Learning

被引:0
|
作者
Li, Zichao [1 ]
Jiang, Xin [1 ]
Shang, Lifeng [1 ]
Li, Hang [2 ]
机构
[1] Huawei Technol, Noahs Ark Lab, Shenzhen, Peoples R China
[2] Toutiao AI Lab, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic generation of paraphrases from a given sentence is an important yet challenging task in natural language processing (NLP). In this paper, we present a deep reinforcement learning approach to paraphrase generation. Specifically, we propose a new framework for the task, which consists of a generator and an evaluator, both of which are learned from data. The generator, built as a sequence-to-sequence learning model, can produce paraphrases given a sentence. The evaluator, constructed as a deep matching model, can judge whether two sentences are paraphrases of each other. The generator is first trained by deep learning and then further fine-tuned by reinforcement learning in which the reward is given by the evaluator. For the learning of the evaluator, we propose two methods based on supervised learning and inverse reinforcement learning respectively, depending on the type of available training data. Experimental results on two datasets demonstrate the proposed models (the generators) can produce more accurate paraphrases and outperform the state-of-the-art methods in paraphrase generation in both automatic evaluation and human evaluation.
引用
收藏
页码:3865 / 3878
页数:14
相关论文
共 50 条
  • [41] From Reinforcement Learning to Deep Reinforcement Learning: An Overview
    Agostinelli, Forest
    Hocquet, Guillaume
    Singh, Sameer
    Baldi, Pierre
    BRAVERMAN READINGS IN MACHINE LEARNING: KEY IDEAS FROM INCEPTION TO CURRENT STATE, 2018, 11100 : 298 - 328
  • [42] Contrastive Representation Learning for Exemplar-Guided Paraphrase Generation ☆
    Yang, Haoran
    Lam, Wai
    Li, Piji
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 4754 - 4761
  • [43] Learning to Selectively Learn for Weakly-supervised Paraphrase Generation
    Ding, Kaize
    Li, Dingcheng
    Li, Alexander Hanbo
    Fan, Xing
    Guo, Chenlei
    Liu, Yang
    Liu, Huan
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 5930 - 5940
  • [44] Transfer Learning in Deep Reinforcement Learning
    Islam, Tariqul
    Abid, Dm. Mehedi Hasan
    Rahman, Tanvir
    Zaman, Zahura
    Mia, Kausar
    Hossain, Ramim
    PROCEEDINGS OF SEVENTH INTERNATIONAL CONGRESS ON INFORMATION AND COMMUNICATION TECHNOLOGY, ICICT 2022, VOL 1, 2023, 447 : 145 - 153
  • [45] Learning to Drive with Deep Reinforcement Learning
    Chukamphaeng, Nut
    Pasupa, Kitsuchart
    Antenreiter, Martin
    Auer, Peter
    2021 13TH INTERNATIONAL CONFERENCE ON KNOWLEDGE AND SMART TECHNOLOGY (KST-2021), 2021, : 147 - 152
  • [46] A Survey on Reinforcement Learning and Deep Reinforcement Learning for Recommender Systems
    Rezaei, Mehrdad
    Tabrizi, Nasseh
    DEEP LEARNING THEORY AND APPLICATIONS, DELTA 2023, 2023, 1875 : 385 - 402
  • [47] Second-Generation Sequencing with Deep Reinforcement Learning for Lung Infection Detection
    Liu, Zhuo
    Zhang, Gerui
    Zhao, Jingyuan
    Yu, Liyan
    Sheng, Junxiu
    Zhang, Na
    Yuan, Hong
    JOURNAL OF HEALTHCARE ENGINEERING, 2020, 2020
  • [48] RTRL: Relation-aware Transformer with Reinforcement Learning for Deep Question Generation
    Zeng, Hongwei
    Wei, Bifan
    Liu, Jun
    KNOWLEDGE-BASED SYSTEMS, 2024, 300
  • [49] Exploratory Policy Generation Methods in On-line Deep Reinforcement Learning: A Survey
    Li, Shilei
    Ye, Qing
    Yuan, Zhimin
    Chen, Yun
    He, Tao
    Fu, Yu
    Jiqiren/Robot, 2024, 46 (06): : 753 - 768
  • [50] Deep reinforcement learning in seat inventory control problem: an action generation approach
    Neda Etebari Alamdari
    Gilles Savard
    Journal of Revenue and Pricing Management, 2021, 20 : 566 - 579