DynamicRetriever: A Pre-trained Model-based IR System Without an Explicit Index

被引:11
|
作者
Zhou, Yu-Jia [1 ]
Yao, Jing [1 ]
Dou, Zhi-Cheng [1 ]
Wu, Ledell [2 ]
Wen, Ji-Rong [1 ]
机构
[1] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing 100872, Peoples R China
[2] Beijing Acad Artificial Intelligence, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Information retrieval (IR); document retrieval; model-based IR; pre-trained language model; differentiable search index;
D O I
10.1007/s11633-022-1373-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Web search provides a promising way for people to obtain information and has been extensively studied. With the surge of deep learning and large-scale pre-training techniques, various neural information retrieval models are proposed, and they have demonstrated the power for improving search (especially, the ranking) quality. All these existing search methods follow a common paradigm, i.e., index-retrieve-rerank, where they first build an index of all documents based on document terms (i.e., sparse inverted index) or representation vectors (i.e., dense vector index), then retrieve and rerank retrieved documents based on the similarity between the query and documents via ranking models. In this paper, we explore a new paradigm of information retrieval without an explicit index but only with a pre-trained model. Instead, all of the knowledge of the documents is encoded into model parameters, which can be regarded as a differentiable indexer and optimized in an end-to-end manner. Specifically, we propose a pre-trained model-based information retrieval (IR) system called DynamicRetriever, which directly returns document identifiers for a given query. Under such a framework, we implement two variants to explore how to train the model from scratch and how to combine the advantages of dense retrieval models. Compared with existing search methods, the model-based IR system parameterizes the traditional static index with a pre-training model, which converts the document semantic mapping into a dynamic and updatable process. Extensive experiments conducted on the public search benchmark Microsoft machine reading comprehension (MS MARCO) verify the effectiveness and potential of our proposed new paradigm for information retrieval.
引用
收藏
页码:276 / 288
页数:13
相关论文
共 50 条
  • [1] PILOT: a pre-trained model-based continual learning toolbox
    Sun, Hai-Long
    Zhou, Da-Wei
    Zhan, De-Chuan
    Ye, Han-Jia
    SCIENCE CHINA-INFORMATION SCIENCES, 2025, 68 (04)
  • [2] PILOT: a pre-trained model-based continual learning toolbox
    HaiLong SUN
    DaWei ZHOU
    DeChuan ZHAN
    HanJia YE
    Science China(Information Sciences), 2025, 68 (04) : 383 - 384
  • [3] Pre-trained Language Model-based Retrieval and Ranking forWeb Search
    Zou, Lixin
    Lu, Weixue
    Liu, Yiding
    Cai, Hengyi
    Chu, Xiaokai
    Ma, Dehong
    Shi, Daiting
    Sun, Yu
    Cheng, Zhicong
    Gu, Simiu
    Wang, Shuaiqiang
    Yin, Dawei
    ACM TRANSACTIONS ON THE WEB, 2023, 17 (01)
  • [4] Pre-Trained Model-Based NFR Classification: Overcoming Limited Data Challenges
    Rahman, Kiramat
    Ghani, Anwar
    Alzahrani, Abdulrahman
    Tariq, Muhammad Usman
    Rahman, Arif Ur
    IEEE ACCESS, 2023, 11 : 81787 - 81802
  • [5] Pre-Trained Model-Based Automated Software Vulnerability Repair: How Far are We?
    Zhang, Quanjun
    Fang, Chunrong
    Yu, Bowen
    Sun, Weisong
    Zhang, Tongke
    Chen, Zhenyu
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2024, 21 (04) : 2507 - 2525
  • [6] Pre-Trained Language Model-Based Deep Learning for Sentiment Classification of Vietnamese Feedback
    Loc, Cu Vinh
    Viet, Truong Xuan
    Viet, Tran Hoang
    Thao, Le Hoang
    Viet, Nguyen Hoang
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS, 2023, 22 (03)
  • [7] Pre-trained Model-based Software Defect Prediction for Edge-cloud Systems
    Kwon, Sunjae
    Lee, Sungu
    Ryu, Duksan
    Baik, Jongmoon
    JOURNAL OF WEB ENGINEERING, 2023, 22 (02): : 255 - 278
  • [8] Expandable Subspace Ensemble for Pre-Trained Model-Based Class-Incremental Learning
    Zhou, Da-Wei
    Sun, Hai-Long
    Ye, Han-Jia
    Zhan, De-Chuan
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 23554 - 23564
  • [9] Pre-trained Model Based Feature Envy Detection
    Ma, Wenhao
    Yu, Yaoxiang
    Ruan, Xiaoming
    Cai, Bo
    2023 IEEE/ACM 20TH INTERNATIONAL CONFERENCE ON MINING SOFTWARE REPOSITORIES, MSR, 2023, : 430 - 440
  • [10] Hyperbolic Pre-Trained Language Model
    Chen, Weize
    Han, Xu
    Lin, Yankai
    He, Kaichen
    Xie, Ruobing
    Zhou, Jie
    Liu, Zhiyuan
    Sun, Maosong
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3101 - 3112