DynamicRetriever: A Pre-trained Model-based IR System Without an Explicit Index

被引:11
|
作者
Zhou, Yu-Jia [1 ]
Yao, Jing [1 ]
Dou, Zhi-Cheng [1 ]
Wu, Ledell [2 ]
Wen, Ji-Rong [1 ]
机构
[1] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing 100872, Peoples R China
[2] Beijing Acad Artificial Intelligence, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Information retrieval (IR); document retrieval; model-based IR; pre-trained language model; differentiable search index;
D O I
10.1007/s11633-022-1373-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Web search provides a promising way for people to obtain information and has been extensively studied. With the surge of deep learning and large-scale pre-training techniques, various neural information retrieval models are proposed, and they have demonstrated the power for improving search (especially, the ranking) quality. All these existing search methods follow a common paradigm, i.e., index-retrieve-rerank, where they first build an index of all documents based on document terms (i.e., sparse inverted index) or representation vectors (i.e., dense vector index), then retrieve and rerank retrieved documents based on the similarity between the query and documents via ranking models. In this paper, we explore a new paradigm of information retrieval without an explicit index but only with a pre-trained model. Instead, all of the knowledge of the documents is encoded into model parameters, which can be regarded as a differentiable indexer and optimized in an end-to-end manner. Specifically, we propose a pre-trained model-based information retrieval (IR) system called DynamicRetriever, which directly returns document identifiers for a given query. Under such a framework, we implement two variants to explore how to train the model from scratch and how to combine the advantages of dense retrieval models. Compared with existing search methods, the model-based IR system parameterizes the traditional static index with a pre-training model, which converts the document semantic mapping into a dynamic and updatable process. Extensive experiments conducted on the public search benchmark Microsoft machine reading comprehension (MS MARCO) verify the effectiveness and potential of our proposed new paradigm for information retrieval.
引用
收藏
页码:276 / 288
页数:13
相关论文
共 50 条
  • [41] A hybrid model in transfer learning based pre-trained model and a scale factor updating
    Jiao, Peng
    Pei, Jing
    2018 JOINT 7TH INTERNATIONAL CONFERENCE ON INFORMATICS, ELECTRONICS & VISION (ICIEV) AND 2018 2ND INTERNATIONAL CONFERENCE ON IMAGING, VISION & PATTERN RECOGNITION (ICIVPR), 2018, : 538 - 543
  • [42] A text restoration model for ancient texts based on pre-trained language model RoBERTa
    Gu, Zhongyu
    Guan, Yanzhi
    Zhang, Shuai
    PROCEEDINGS OF 2024 4TH INTERNATIONAL CONFERENCE ON INTERNET OF THINGS AND MACHINE LEARNING, IOTML 2024, 2024, : 96 - 102
  • [43] EventBERT: A Pre-Trained Model for Event Correlation Reasoning
    Zhou, Yucheng
    Geng, Xiubo
    Shen, Tao
    Long, Guodong
    Jiang, Daxin
    PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 850 - 859
  • [44] CODEFUSION: A Pre-trained Diffusion Model for Code Generation
    Singh, Mukul
    Cambronero, Jose
    Gulwani, Sumit
    Le, Vu
    Negreanu, Carina
    Verbruggen, Gust
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 11697 - 11708
  • [45] Generating Empathetic Responses with a Pre-trained Conversational Model
    Beredo, Jackylyn
    Bautista, Carlo Migel
    Cordel, Macario
    Ong, Ethel
    TEXT, SPEECH, AND DIALOGUE, TSD 2021, 2021, 12848 : 147 - 158
  • [46] PTMA: Pre-trained Model Adaptation for Transfer Learning
    Li, Xiao
    Yan, Junkai
    Jiang, Jianjian
    Zheng, Wei-Shi
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2024, 2024, 14884 : 176 - 188
  • [47] Pre-trained Language Model Representations for Language Generation
    Edunov, Sergey
    Baevski, Alexei
    Auli, Michael
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 4052 - 4059
  • [48] A Pre-trained Knowledge Tracing Model with Limited Data
    Yue, Wenli
    Su, Wei
    Liu, Lei
    Cai, Chuan
    Yuan, Yongna
    Jia, Zhongfeng
    Liu, Jiamin
    Xie, Wenjian
    DATABASE AND EXPERT SYSTEMS APPLICATIONS, PT I, DEXA 2024, 2024, 14910 : 163 - 178
  • [49] ViDeBERTa: A powerful pre-trained language model for Vietnamese
    Tran, Cong Dao
    Pham, Nhut Huy
    Nguyen, Anh
    Hy, Truong Son
    Vu, Tu
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1071 - 1078
  • [50] CodeBERT: A Pre-Trained Model for Programming and Natural Languages
    Feng, Zhangyin
    Guo, Daya
    Tang, Duyu
    Duan, Nan
    Feng, Xiaocheng
    Gong, Ming
    Shou, Linjun
    Qin, Bing
    Liu, Ting
    Jiang, Daxin
    Zhou, Ming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1536 - 1547