KnowPrefix-Tuning: A Two-Stage Prefix-Tuning Framework for Knowledge-Grounded Dialogue Generation

被引:1
|
作者
Bai, Jiaqi [1 ,2 ]
Yan, Zhao [3 ]
Yang, Ze [2 ]
Yang, Jian [2 ]
Liang, Xinnian [2 ]
Guo, Hongcheng [2 ]
Li, Zhoujun [1 ,2 ]
机构
[1] Beihang Univ, Sch Cyber Sci & Technol, Beijing, Peoples R China
[2] Beihang Univ, State Key Lab Software Dev Environm, Beijing, Peoples R China
[3] Tencent Cloud AI, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Dialogue generation; Parameter-efficient fine-tuning; Knowledge-grounded dialogue; Pre-trained language models;
D O I
10.1007/978-3-031-43415-0_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing knowledge-grounded conversation systems generate responses typically in a retrieve-then-generate manner. They require a large knowledge base and a strong knowledge retrieval component, which is time- and resource-consuming. In this paper, we address the challenge by leveraging the inherent knowledge encoded in the pre-trained language models (PLMs). We propose Knowledgeable Prefix Tuning (KnowPrefix-Tuning), a two-stage tuning framework, bypassing the retrieval process in a knowledge-grounded conversation system by injecting prior knowledge into the lightweight knowledge prefix. The knowledge prefix is a sequence of continuous knowledge-specific vectors that can be learned during training. In addition, we propose a novel interactive re-parameterization mechanism that allows the prefix to interact fully with the PLM during the optimization of response generation. Experimental results demonstrate that KnowPrefix-Tuning outperforms fine-tuning and other lightweight tuning approaches, and performs comparably with strong retrieval-based baselines while being 3x faster during inference (The code is available at https://github.com/fantast4ever/KnowPrefix-Tuning.)
引用
收藏
页码:525 / 542
页数:18
相关论文
共 8 条
  • [1] PICKD: In-Situ Prompt Tuning for Knowledge-Grounded Dialogue Generation
    Sarkar, Rajdeep
    Goswami, Koustava
    Arcan, Mihael
    McCrae, John
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT IV, 2023, 13938 : 124 - 136
  • [2] A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation
    Liu, Shilei
    Zhao, Xiaofeng
    Li, Bochao
    Ren, Feiliang
    Zhang, Longhui
    Yin, Shujuan
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 2262 - 2272
  • [3] Training Two-Stage Knowledge-Grounded Dialogues with Attention Feedback
    Li, Zhen
    Feng, Jiazhan
    Tao, Chongyang
    Zhao, Dongyan
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 473 - 484
  • [4] XDAI: A Tuning-free Framework for Exploiting Pre-trained Language Models in Knowledge Grounded Dialogue Generation
    Yu, Jifan
    Zhang, Xiaohan
    Xu, Yifan
    Lei, Xuanyu
    Guan, Xinyu
    Zhang, Jing
    Hou, Lei
    Li, Juanzi
    Tang, Jie
    [J]. PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 4422 - 4432
  • [5] A Knowledge-enhanced Two-stage Generative Framework for Medical Dialogue Information Extraction
    Zefa Hu
    Ziyi Ni
    Jing Shi
    Shuang Xu
    Bo Xu
    [J]. Machine Intelligence Research, 2024, 21 : 153 - 168
  • [6] A Knowledge-enhanced Two-stage Generative Framework for Medical Dialogue Information Extraction
    Hu, Zefa
    Ni, Ziyi
    Shi, Jing
    Xu, Shuang
    Xu, Bo
    [J]. MACHINE INTELLIGENCE RESEARCH, 2024, 21 (01) : 153 - 168
  • [7] Improving knowledge-based dialogue generation through two-stage knowledge selection and knowledge selection-guided pointer network
    Liu, Mengjuan
    Zhao, Pei
    Liu, Jiang
    Zhuang, Yulin
    Yang, Yunfan
    [J]. JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2022, 59 (03) : 591 - 611
  • [8] Improving knowledge-based dialogue generation through two-stage knowledge selection and knowledge selection-guided pointer network
    Mengjuan Liu
    Pei Zhao
    Jiang Liu
    Yulin Zhuang
    Yunfan Yang
    [J]. Journal of Intelligent Information Systems, 2022, 59 : 591 - 611