FactGen: Faithful Text Generation by Factuality-aware Pre-training and Contrastive Ranking Fine-tuning

被引:0
|
作者
Lan, Zhibin [1 ,2 ]
Li, Wei [3 ]
Su, Jinsong [1 ,2 ]
Xiao, Xinyan [3 ]
Liu, Jiachen [3 ]
Wu, Wenhao [3 ]
Lyu, Yajuan [3 ]
机构
[1] Xiamen Univ, Sch Informat, Xiamen, Peoples R China
[2] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
[3] Baidu, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Conditional text generation is supposed to generate a fluent and coherent target text that is faithful to the source text. Although pre-trained models have achieved promising results, they still suffer from the crucial factuality problem. To deal with this issue, we propose a factuality-aware pretraining-finetuning framework named FactGen, which fully considers factuality during two training stages. Specifically, at the pre-training stage, we utilize a natural language inference model to construct target texts that are entailed by the source texts, resulting in a more factually consistent pre-training objective. Then, during the fine-tuning stage, we further introduce a contrastive ranking loss to encourage the model to generate factually consistent text with higher probability. Extensive experiments on three conditional text generation tasks demonstrate the effectiveness and generality of our training framework.
引用
收藏
页码:1281 / 1303
页数:23
相关论文
共 50 条
  • [41] Evaluation of Dataset Selection for Pre-Training and Fine-Tuning Transformer Language Models for Clinical Question Answering
    Soni, Sarvesh
    Roberts, Kirk
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 5532 - 5538
  • [42] From pre-training to fine-tuning: An in-depth analysis of Large Language Models in the biomedical domain
    Bonfigli, Agnese
    Bacco, Luca
    Merone, Mario
    Dell'Orletta, Felice
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 157
  • [43] Pre-training using pseudo images and fine-tuning using real images for nighttime traffic Sign Detection
    Yamamoto M.
    Ohashi G.
    IEEJ Transactions on Electronics, Information and Systems, 2021, 141 (09) : 969 - 976
  • [44] Training Deep Spiking Convolutional Neural Networks With STDP-Based Unsupervised Pre-training Followed by Supervised Fine-Tuning
    Lee, Chankyu
    Panda, Priyadarshini
    Srinivasan, Gopalakrishnan
    Roy, Kaushik
    FRONTIERS IN NEUROSCIENCE, 2018, 12
  • [45] Breaking the Barrier Between Pre-training and Fine-tuning: A Hybrid Prompting Model for Knowledge-Based VQA
    Sun, Zhongfan
    Hu, Yongli
    Gao, Qingqing
    Jiang, Huajie
    Gao, Junbin
    Sun, Yanfeng
    Yin, Baocai
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4065 - 4073
  • [46] Adaptive Pre-Training and Collaborative Fine-Tuning: A Win-Win Strategy to Improve Review Analysis Tasks
    Mao, Qianren
    Li, Jianxin
    Lin, Chenghua
    Chen, Congwen
    Peng, Hao
    Wang, Lihong
    Yu, Philip S.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 622 - 634
  • [47] Pre-training and Fine-tuning Neural Topic Model: A Simple yet Effective Approach to Incorporating External Knowledge
    Zhang, Linhai
    Hu, Xumeng
    Wang, Boyu
    Zhou, Deyu
    Zhang, Qian-Wen
    Cao, Yunbo
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5980 - 5989
  • [48] APF-GAN: Exploring asymmetric pre-training and fine-tuning strategy for conditional generative adversarial network
    Yuxuan Li
    Lingfeng Yang
    Xiang Li
    Computational Visual Media, 2024, 10 : 187 - 192
  • [49] Fine-tuning Pre-trained Language Models for Few-shot Intent Detection: Supervised Pre-training and Isotropization
    Zhang, Haode
    Liang, Haowen
    Zhang, Yuwei
    Zhan, Liming
    Wu, Xiao-Ming
    Lu, Xiaolei
    Lam, Albert Y. S.
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 532 - 542
  • [50] APF-GAN: Exploring asymmetric pre-training and fine-tuning strategy for conditional generative adversarial network
    Li, Yuxuan
    Yang, Lingfeng
    Li, Xiang
    COMPUTATIONAL VISUAL MEDIA, 2024, 10 (01): : 187 - 192