Paraphrase Generation with Deep Reinforcement Learning

被引:0
|
作者
Li, Zichao [1 ]
Jiang, Xin [1 ]
Shang, Lifeng [1 ]
Li, Hang [2 ]
机构
[1] Huawei Technol, Noahs Ark Lab, Shenzhen, Peoples R China
[2] Toutiao AI Lab, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Automatic generation of paraphrases from a given sentence is an important yet challenging task in natural language processing (NLP). In this paper, we present a deep reinforcement learning approach to paraphrase generation. Specifically, we propose a new framework for the task, which consists of a generator and an evaluator, both of which are learned from data. The generator, built as a sequence-to-sequence learning model, can produce paraphrases given a sentence. The evaluator, constructed as a deep matching model, can judge whether two sentences are paraphrases of each other. The generator is first trained by deep learning and then further fine-tuned by reinforcement learning in which the reward is given by the evaluator. For the learning of the evaluator, we propose two methods based on supervised learning and inverse reinforcement learning respectively, depending on the type of available training data. Experimental results on two datasets demonstrate the proposed models (the generators) can produce more accurate paraphrases and outperform the state-of-the-art methods in paraphrase generation in both automatic evaluation and human evaluation.
引用
收藏
页码:3865 / 3878
页数:14
相关论文
共 50 条
  • [31] Edge Generation Scheduling for DAG Tasks Using Deep Reinforcement Learning
    Sun, Binqi
    Theile, Mirco
    Qin, Ziyuan
    Bernardini, Daniele
    Roy, Debayan
    Bastoni, Andrea
    Caccamo, Marco
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (04) : 1034 - 1047
  • [32] Efficient and robust entanglement generation with deep reinforcement learning for quantum metrology
    Qiu, Yuxiang
    Zhuang, Min
    Huang, Jiahao
    Lee, Chaohong
    NEW JOURNAL OF PHYSICS, 2022, 24 (08):
  • [33] Hierarchical Gait Generation for Modular Robots Using Deep Reinforcement Learning
    Wang, Jiayu
    Hu, Chuxiong
    Zhu, Yu
    2021 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS (ICM), 2021,
  • [34] Conformer-RL: A deep reinforcement learning library for conformer generation
    Jiang, Runxuan
    Gogineni, Tarun
    Kammeraad, Joshua
    He, Yifei
    Tewari, Ambuj
    Zimmerman, Paul M.
    JOURNAL OF COMPUTATIONAL CHEMISTRY, 2022, 43 (27) : 1880 - 1886
  • [35] Reinforcement learning-driven deep question generation with rich semantics
    Guan, Menghong
    Mondal, Subrota Kumar
    Dai, Hong-Ning
    Bao, Haiyong
    INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (02)
  • [36] Animation generation for object transportation with a rope using deep reinforcement learning
    Wong, Sai-Keung
    Wei, Xu-Tao
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2023, 34 (3-4)
  • [37] Deep reinforcement learning control of white-light continuum generation
    Valensise, Carlo M.
    Giuseppi, Alessandro
    Cerullo, Giulio
    Polli, Dario
    OPTICA, 2021, 8 (02): : 239 - 242
  • [38] DEEP REINFORCEMENT LEARNING-BASED AUTOMATIC TEST PATTERN GENERATION
    Li, Wenxing
    Lyu, Hongqin
    Liang, Shengwen
    Liu, Zizhen
    Lin, Ning
    Wang, Zhongrui
    Tian, Pengyu
    Wang, Tiancheng
    Li, Huawei
    CONFERENCE OF SCIENCE & TECHNOLOGY FOR INTEGRATED CIRCUITS, 2024 CSTIC, 2024,
  • [39] Automated Gait Generation for Simulated Bodies using Deep Reinforcement Learning
    Ananthakrishnan, Abhishek
    Kanakiya, Vatsal
    Ved, Dipen
    Sharma, Grishma
    PROCEEDINGS OF THE 2018 SECOND INTERNATIONAL CONFERENCE ON INVENTIVE COMMUNICATION AND COMPUTATIONAL TECHNOLOGIES (ICICCT), 2018, : 90 - 95
  • [40] Automatic Generation Control Based on Deep Reinforcement Learning With Exploration Awareness
    Xi L.
    Yu L.
    Fu Y.
    Huang Y.
    Chen X.
    Kang S.
    Zhongguo Dianji Gongcheng Xuebao/Proceedings of the Chinese Society of Electrical Engineering, 2019, 39 (14): : 4150 - 4161