A Survey of Controllable Text Generation Using Transformer-based Pre-trained Language Models

被引:23
|
作者
Zhang, Hanqing [1 ]
Song, Haolin [1 ]
Li, Shaoyu [1 ]
Zhou, Ming [2 ]
Song, Dawei [1 ]
机构
[1] Beijing Inst Technol, 5 South St, Beijing 100081, Peoples R China
[2] Langboat Technol, 52 Beisihuan West Rd, Beijing 100081, Peoples R China
关键词
Controllable text generation; pre-trained language models; Transformer; controllability; systematic review;
D O I
10.1145/3617680
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Controllable Text Generation (CTG) is an emerging area in the field of natural language generation (NLG). It is regarded as crucial for the development of advanced text generation technologies that better meet the specific constraints in practical applications. In recent years, methods using large-scale pre-trained language models (PLMs), in particular the widely used Transformer-based PLMs, have become a new paradigm of NLG, allowing generation of more diverse and fluent text. However, due to the limited level of interpretability of deep neural networks, the controllability of these methods needs to be guaranteed. To this end, controllable text generation using Transformer-based PLMs has become a rapidly growing yet challenging new research hotspot. A diverse range of approaches have emerged in the past 3 to 4 years, targeting different CTG tasks that require different types of controlled constraints. In this article, we present a systematic critical review on the common tasks, main approaches, and evaluation methods in this area. Finally, we discuss the challenges that the field is facing, and put forward various promising future directions. To the best of our knowledge, this is the first survey article to summarize the state-of-the-art CTG techniques from the perspective of Transformer-based PLMs. We hope it can help researchers and practitioners in the related fields to quickly track the academic and technological frontier, providing them with a landscape of the area and a roadmap for future research.
引用
收藏
页数:37
相关论文
共 50 条
  • [41] Text Detoxification using Large Pre-trained Neural Models
    Dale, David
    Voronov, Anton
    Dementieva, Daryna
    Logacheva, Varvara
    Kozlova, Olga
    Semenov, Nikita
    Panchenko, Alexander
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 7979 - 7996
  • [42] CPT: a pre-trained unbalanced transformer for both Chinese language understanding and generation
    Shao, Yunfan
    Geng, Zhichao
    Liu, Yitao
    Dai, Junqi
    Yan, Hang
    Yang, Fei
    Li, Zhe
    Bao, Hujun
    Qiu, Xipeng
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (05)
  • [43] Effective test generation using pre-trained Large Language Models and mutation testing
    Dakhel, Arghavan Moradi
    Nikanjam, Amin
    Majdinasab, Vahid
    Khomh, Foutse
    Desmarais, Michel C.
    [J]. INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 171
  • [44] An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language Models
    Meade, Nicholas
    Poole-Dayan, Elinor
    Reddy, Siva
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1878 - 1898
  • [45] Question Answering based Clinical Text Structuring Using Pre-trained Language Model
    Qiu, Jiahui
    Zhou, Yangming
    Ma, Zhiyuan
    Ruan, Tong
    Liu, Jinlin
    Sun, Jing
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2019, : 1596 - 1600
  • [46] ProtTrans-Glutar: Incorporating Features From Pre-trained Transformer-Based Models for Predicting Glutarylation Sites
    Indriani, Fatma
    Mahmudah, Kunti Robiatul
    Purnama, Bedy
    Satou, Kenji
    [J]. FRONTIERS IN GENETICS, 2022, 13
  • [47] Pre-trained Language Model Representations for Language Generation
    Edunov, Sergey
    Baevski, Alexei
    Auli, Michael
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 4052 - 4059
  • [48] Knowledge-Grounded Dialogue Generation with Pre-trained Language Models
    Zhao, Xueliang
    Wu, Wei
    Xu, Can
    Tao, Chongyang
    Zhao, Dongyan
    Yan, Rui
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3377 - 3390
  • [49] Automatic Title Generation for Learning Resources and Pathways with Pre-trained Transformer Models
    Mishra, Prakhar
    Diwan, Chaitali
    Srinivasa, Srinath
    Srinivasaraghavan, G.
    [J]. INTERNATIONAL JOURNAL OF SEMANTIC COMPUTING, 2021, 15 (04) : 487 - 510
  • [50] μBERT: Mutation Testing using Pre-Trained Language Models
    Degiovanni, Renzo
    Papadakis, Mike
    [J]. 2022 IEEE 15TH INTERNATIONAL CONFERENCE ON SOFTWARE TESTING, VERIFICATION AND VALIDATION WORKSHOPS (ICSTW 2022), 2022, : 160 - 169