POINTER: Constrained Progressive Text Generation via Insertion-based Generative Pre-training

被引:0
|
作者
Zhang, Yizhe [1 ]
Wang, Guoyin [2 ]
Li, Chunyuan [1 ]
Gan, Zhe [1 ]
Brockett, Chris [1 ]
Dolan, Bill [1 ]
机构
[1] Microsoft Res, Redmond, WA 98052 USA
[2] Amazon Alexa AI, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large-scale pre-trained language models, such as BERT and GPT-2, have achieved excellent performance in language representation learning and free-form text generation. However, these models cannot be directly employed to generate text under specified lexical constraints. To address this challenge, we present POINTER', a simple yet novel insertion-based approach for hard-constrained text generation. The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner. This procedure is recursively applied until a sequence is completed. The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable. We pre-train our model with the proposed progressive insertion-based objective on a 12GB Wikipedia dataset, and finetune it on downstream hard-constrained generation tasks. Non-autoregressive decoding yields an empirically logarithmic time complexity during inference time. Experimental results on both News and Yelp datasets demonstrate that POINTER achieves state-of-the-art performance on constrained text generation. We released the pre-trained models and the source code to facilitate future research(2).
引用
收藏
页码:8649 / 8670
页数:22
相关论文
共 50 条
  • [1] ENCONTER: Entity Constrained Progressive Sequence Generation via Insertion-based Transformer
    Hsieh, Lee-Hsun
    Lee, Yang-Yin
    Lim, Ee-Peng
    [J]. 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 3590 - 3599
  • [2] RAR: Recombination and augmented replacement method for insertion-based lexically constrained text generation
    Kang, Fengrui
    Huang, Xianying
    Li, Bingyu
    [J]. NEUROCOMPUTING, 2024, 597
  • [3] MolXPT: Wrapping Molecules with Text for Generative Pre-training
    Liu, Zequn
    Zhang, Wei
    Xia, Yingce
    Wu, Lijun
    Xie, Shufang
    Qin, Tao
    Zhang, Ming
    Liu, Tie-Yan
    [J]. 61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1606 - 1616
  • [4] Self-attention Based Text Matching Model with Generative Pre-training
    Zhang, Xiaolin
    Lei, Fengpei
    Yu, Shengji
    [J]. 2021 IEEE INTL CONF ON DEPENDABLE, AUTONOMIC AND SECURE COMPUTING, INTL CONF ON PERVASIVE INTELLIGENCE AND COMPUTING, INTL CONF ON CLOUD AND BIG DATA COMPUTING, INTL CONF ON CYBER SCIENCE AND TECHNOLOGY CONGRESS DASC/PICOM/CBDCOM/CYBERSCITECH 2021, 2021, : 84 - 91
  • [5] Learning Visual Prior via Generative Pre-Training
    Xie, Jinheng
    Ye, Kai
    Li, Yudong
    Li, Yuexiang
    Lin, Kevin Qinghong
    Zheng, Yefeng
    Shen, Linlin
    Shou, Mike Zheng
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Denoising based Sequence-to-Sequence Pre-training for Text Generation
    Wang, Liang
    Zhao, Wei
    Jia, Ruoyu
    Li, Sujian
    Liu, Jingming
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 4003 - 4015
  • [7] RecipeGPT: Generative Pre-training Based Cooking Recipe Generation and Evaluation System
    Lee, Helena H.
    Shu, Ke
    Achananuparp, Palakorn
    Prasetyo, Philips Kokoh
    Liu, Yue
    Lim, Ee-Peng
    Varshney, Lav R.
    [J]. WWW'20: COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2020, 2020, : 181 - 184
  • [8] INSNET: An Efficient, Flexible, and Performant Insertion-based Text Generation Model
    Lu, Sidi
    Meng, Tao
    Peng, Nanyun
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [9] HG-News: News Headline Generation Based on a Generative Pre-Training Model
    Li, Ping
    Yu, Jiong
    Chen, Jiaying
    Guo, Binglei
    [J]. IEEE ACCESS, 2021, 9 : 110039 - 110046
  • [10] Generative Pre-training for Paraphrase Generation by Representing and Predicting Spans in Exemplars
    Bui, Tien-Cuong
    Le, Van-Duc
    To, Hai-Thien
    Cha, Sang Kyun
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP 2021), 2021, : 83 - 90