Improving Dialog System Grounded with Unstructured Knowledge by Domain Adaptive Pre-Training and Post-Ranking

被引:1
|
作者
Zeng, Zhi [1 ]
Lam, Kam-Yiu [1 ]
Chow, Chi-Yin [1 ]
Li, Ning [1 ]
机构
[1] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
关键词
Domain adaptive training; dynamic masking probability; post-ranking; knowledge selection; knowledge-grounded generation;
D O I
10.1142/S0219843621500195
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Linguistic intelligence and the ability to converse with human are important and indispensable parts of humanoid robots. One of the most challenging tasks in knowledge-grounded task-oriented dialog systems (KTDS) is the knowledge selection task, which aims to find the proper knowledge snippets to respond to user dialog requests. In this paper, we first propose domain adapted-BERT (DA-BERT) which employs pre-trained bidirectional encoder representations from transformers (BERT) with domain adaptive training and dynamic masking probability for knowledge selection in KTDS. Domain adaptive training can minimize the domain gap between the general text data that BERT is pre-trained on and the dialog-knowledge joint data while dynamic masking probability enhances the training in an easy-to-hard manner. After knowledge selection, the next task in KTDS is knowledge-grounded generation. To improve the performance in knowledge-grounded generation, we propose GPT-PR to employ post-ranking on the generator's outputs. Post-ranking eliminates the possibility of generating hallucination response by a large portion during the sampling-based decoding process and thus can improve the quality of the generated response. Experimental results on the benchmark dataset show that our proposed pre-training and post-ranking methods, DA-BERT and GPT-PR, respectively, outperform the state-of-the-art models with large margins across all the evaluation metrics. Moreover, in the experiments, we also analyze the bad cases of DA-BERT and GPT-PR and do visualizations to facilitate further research in this direction.
引用
收藏
页数:25
相关论文
共 32 条
  • [1] GLM-Dialog: Noise-tolerant Pre-training for Knowledge-grounded Dialogue Generation
    Zhang, Jing
    Zhang, Xiaokang
    Zhang-Li, Daniel
    Yu, Jifan
    Yao, Zijun
    Ma, Zeyao
    Xu, Yiqi
    Wang, Haohua
    Zhang, Xiaohan
    Lin, Nianyi
    Lu, Sunrui
    Li, Juanzi
    Tang, Jie
    [J]. PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5564 - 5575
  • [2] Improving Linguistic Bias Detection in Wikipedia using Cross-Domain Adaptive Pre-Training
    Madanagopal, Karthic
    Caverlee, James
    [J]. COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2022, WWW 2022 COMPANION, 2022, : 1301 - 1309
  • [3] KGPT: Knowledge-Grounded Pre-Training for Data-to-Text Generation
    Chen, Wenhu
    Su, Yu
    Yan, Xifeng
    Wang, William Yang
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 8635 - 8648
  • [4] Improving Knowledge Tracing via Pre-training Question Embeddings
    Liu, Yunfei
    Yang, Yang
    Chen, Xianyu
    Shen, Jian
    Zhang, Haifeng
    Yu, Yong
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1577 - 1583
  • [5] Improving Knowledge Tracing via Pre-training Question Embeddings
    Liu, Yunfei
    Yang, Yang
    Chen, Xianyu
    Shen, Jian
    Zhang, Haifeng
    Yu, Yong
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 1556 - 1562
  • [6] Grounded Entity-Landmark Adaptive Pre-training for Vision-and-Language Navigation
    Cui, Yibo
    Xie, Liang
    Zhang, Yakun
    Zhang, Meishan
    Yan, Ye
    Yin, Erwei
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 12009 - 12019
  • [7] K-DLM: A Domain-Adaptive Language Model Pre-Training Framework with Knowledge Graph
    Zou, Jiaxin
    Xie, Zuotong
    Chen, Junhua
    Hou, Jiawei
    Yan, Qiang
    Zheng, Hai-Tao
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT IV, 2023, 14257 : 447 - 459
  • [8] Improving Knowledge Graph Representation Learning by Structure Contextual Pre-training
    Ye, Ganqiang
    Zhang, Wen
    Bi, Zhen
    Wong, Chi Man
    Chen, Hui
    Chen, Huajun
    [J]. PROCEEDINGS OF THE 10TH INTERNATIONAL JOINT CONFERENCE ON KNOWLEDGE GRAPHS (IJCKG 2021), 2021, : 151 - 155
  • [9] A Pre-training Strategy for Zero-Resource Response Selection in Knowledge-Grounded Conversations
    Tao, Chongyang
    Chen, Changyu
    Feng, Jiazhan
    Wen, Ji-Rong
    Yan, Rui
    [J]. 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4446 - 4457
  • [10] A knowledge-guided pre-training framework for improving molecular representation learning
    Li, Han
    Zhang, Ruotian
    Min, Yaosen
    Ma, Dacheng
    Zhao, Dan
    Zeng, Jianyang
    [J]. NATURE COMMUNICATIONS, 2023, 14 (01)