A Survey of Controllable Text Generation Using Transformer-based Pre-trained Language Models

被引:23
|
作者
Zhang, Hanqing [1 ]
Song, Haolin [1 ]
Li, Shaoyu [1 ]
Zhou, Ming [2 ]
Song, Dawei [1 ]
机构
[1] Beijing Inst Technol, 5 South St, Beijing 100081, Peoples R China
[2] Langboat Technol, 52 Beisihuan West Rd, Beijing 100081, Peoples R China
关键词
Controllable text generation; pre-trained language models; Transformer; controllability; systematic review;
D O I
10.1145/3617680
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Controllable Text Generation (CTG) is an emerging area in the field of natural language generation (NLG). It is regarded as crucial for the development of advanced text generation technologies that better meet the specific constraints in practical applications. In recent years, methods using large-scale pre-trained language models (PLMs), in particular the widely used Transformer-based PLMs, have become a new paradigm of NLG, allowing generation of more diverse and fluent text. However, due to the limited level of interpretability of deep neural networks, the controllability of these methods needs to be guaranteed. To this end, controllable text generation using Transformer-based PLMs has become a rapidly growing yet challenging new research hotspot. A diverse range of approaches have emerged in the past 3 to 4 years, targeting different CTG tasks that require different types of controlled constraints. In this article, we present a systematic critical review on the common tasks, main approaches, and evaluation methods in this area. Finally, we discuss the challenges that the field is facing, and put forward various promising future directions. To the best of our knowledge, this is the first survey article to summarize the state-of-the-art CTG techniques from the perspective of Transformer-based PLMs. We hope it can help researchers and practitioners in the related fields to quickly track the academic and technological frontier, providing them with a landscape of the area and a roadmap for future research.
引用
收藏
页数:37
相关论文
共 50 条
  • [21] Pre-trained models for natural language processing: A survey
    QIU XiPeng
    SUN TianXiang
    XU YiGe
    SHAO YunFan
    DAI Ning
    HUANG XuanJing
    [J]. Science China(Technological Sciences), 2020, (10) - 1897
  • [22] Pre-trained models for natural language processing: A survey
    XiPeng Qiu
    TianXiang Sun
    YiGe Xu
    YunFan Shao
    Ning Dai
    XuanJing Huang
    [J]. Science China Technological Sciences, 2020, 63 : 1872 - 1897
  • [23] BioGPT: generative pre-trained transformer for biomedical text generation and mining
    Luo, Renqian
    Sun, Liai
    Xia, Yingce
    Qin, Tao
    Zhang, Sheng
    Poon, Hoifung
    Liu, Tie-Yan
    [J]. BRIEFINGS IN BIOINFORMATICS, 2022, 23 (06)
  • [24] Automatic text summarization using transformer-based language models
    Rao, Ritika
    Sharma, Sourabh
    Malik, Nitin
    [J]. INTERNATIONAL JOURNAL OF SYSTEM ASSURANCE ENGINEERING AND MANAGEMENT, 2024, 15 (06) : 2599 - 2605
  • [25] Leveraging pre-trained language models for code generation
    Soliman, Ahmed
    Shaheen, Samir
    Hadhoud, Mayada
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (03) : 3955 - 3980
  • [26] Towards a Transformer-Based Pre-trained Model for IoT Traffic Classification
    Bazaluk, Bruna
    Hamdan, Mosab
    Ghaleb, Mustafa
    Gismalla, Mohammed S. M.
    da Silva, Flavio S. Correa
    Batista, Daniel Macedo
    [J]. PROCEEDINGS OF 2024 IEEE/IFIP NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM, NOMS 2024, 2024,
  • [27] A Robust Approach to Fine-tune Pre-trained Transformer-based models for Text Summarization through Latent Space Compression
    Falaki, Ala Alam
    Gras, Robin
    [J]. 2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 160 - 167
  • [28] Text clustering based on pre-trained models and autoencoders
    Xu, Qiang
    Gu, Hao
    Ji, ShengWei
    [J]. FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2024, 17
  • [29] Extremely Low Resource Text simplification with Pre-trained Transformer Language Model
    Maruyama, Takumi
    Yamamoto, Kazuhide
    [J]. PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2019, : 53 - 58
  • [30] Pre-trained Language Models in Biomedical Domain: A Systematic Survey
    Wang, Benyou
    Xie, Qianqian
    Pei, Jiahuan
    Chen, Zhihong
    Tiwari, Prayag
    Li, Zhao
    Fu, Jie
    [J]. ACM COMPUTING SURVEYS, 2024, 56 (03)