Distilling the Knowledge of Large-scale Generative Models into Retrieval Models for Efficient Open-domain Conversation

被引:0
|
作者
Kim, Beomsu [1 ]
Seo, Seokjun [1 ]
Han, Seungju [1 ]
Erdenee, Enkhbayar [1 ]
Chang, Buru [1 ]
机构
[1] Hyperconnect, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the remarkable performance of largescale generative models in open-domain conversation, they are known to be less practical for building real-time conversation systems due to high latency. On the other hand, retrieval models could return responses with much lower latency but show inferior performance to the large-scale generative models since the conversation quality is bounded by the pre-defined response set. To take advantage of both approaches, we propose a new training method called G2R (Generative-toRetrieval distillation) that preserves the efficiency of a retrieval model while leveraging the conversational ability of a large-scale generative model by infusing the knowledge of the generative model into the retrieval model. G2R consists of two distinct techniques of distillation: the data-level G2R augments the dialogue dataset with additional responses generated by the large-scale generative model, and the model-level G2R transfers the response quality score assessed by the generative model to the score of the retrieval model by the knowledge distillation loss. Through extensive experiments including human evaluation, we demonstrate that our retrieval-based conversation system trained with G2R shows a substantially improved performance compared to the baseline retrieval model while showing significantly lower inference latency than the largescale generative models.
引用
收藏
页码:3357 / 3373
页数:17
相关论文
共 50 条
  • [1] TwinBERT: Distilling Knowledge to Twin-Structured Compressed BERT Models for Large-Scale Retrieval
    Lu, Wenhao
    Jiao, Jian
    Zhang, Ruofei
    [J]. CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 2645 - 2652
  • [2] Building a Role Specified Open-Domain Dialogue System Leveraging Large-Scale Language Models
    Bae, Sanghwan
    Kwak, Donghyun
    Kim, Sungdong
    Ham, Donghoon
    Kang, Soyoung
    Lee, Sang-Woo
    Park, Woomyoung
    [J]. NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 2128 - 2150
  • [3] Generative commonsense knowledge subgraph retrieval for open-domain dialogue response generation
    Wu, Sixing
    Yu, Jiong
    Chen, Jiahao
    Zhou, Wei
    [J]. NEURAL NETWORKS, 2024, 180
  • [4] Towards Empathetic Open-domain Conversation Models: a New Benchmark and Dataset
    Rashkin, Hannah
    Smith, Eric Michael
    Li, Margaret
    Boureau, Y-Lan
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5370 - 5381
  • [5] Evaluating Open-Domain Question Answering in the Era of Large Language Models
    Kamalloo, Ehsan
    Dziri, Nouha
    Clarke, Charles L. A.
    Rafiei, Davood
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 5591 - 5606
  • [6] Open-Domain Question Answering over Tables with Large Language Models
    Liang, Xinyi
    Hu, Rui
    Liu, Yu
    Zhu, Konglin
    [J]. ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT XII, ICIC 2024, 2024, 14873 : 347 - 358
  • [7] GENERATIVE MODELS FOR LARGE-SCALE SIMULATIONS OF CONNECTOME DEVELOPMENT
    Brooks, Skylar J.
    Stamoulis, Catherine
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [8] Will Large-scale Generative Models Corrupt Future Datasets?
    Hataya, Ryuichiro
    Bao, Han
    Arai, Hiromi
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 20498 - 20508
  • [9] Generative models and abstractions for large-scale neuroanatomy datasets
    Rolnick, David
    Dyer, Eva L.
    [J]. CURRENT OPINION IN NEUROBIOLOGY, 2019, 55 : 112 - 120
  • [10] Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering
    Izacard, Gautier
    Grave, Edouard
    [J]. 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 874 - 880