Pre-training Methods in Information Retrieval

被引:21
|
作者
Fan, Yixing [1 ]
Xie, Xiaohui [2 ]
Cai, Yinqiong [1 ]
Chen, Jia [2 ]
Ma, Xinyu [1 ]
Li, Xiangsheng [2 ]
Zhang, Ruqing [1 ]
Guo, Jiafeng [1 ]
机构
[1] Chinese Acad Sci, ICT, Beijing, Peoples R China
[2] Tsinghua Univ, Beijing, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
MODELS;
D O I
10.1561/1500000100
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The core of information retrieval (IR) is to identify relevant information from large-scale resources and return it as a ranked list to respond to user's information need. In recent years, the resurgence of deep learning has greatly advanced this field and leads to a hot topic named NeuIR (i.e., neural information retrieval), especially the paradigm of pre-training methods (PTMs). Owing to sophisticated pre-training objectives and huge model size, pre-trained models can learn universal language representations from massive textual data, which are beneficial to the ranking task of IR. Recently, a large number of works, which are dedicated to the application of PTMs in IR, have been introduced to promote the retrieval performance. Considering the rapid progress of this direction, this survey aims to provide a systematic review of pre-training methods in IR. To be specific, we present an overview of PTMs applied in different components of an IR system, including the retrieval component, the re-ranking component, and other components. In addition, we also introduce PTMs specifically designed for IR, and summarize available datasets as well as benchmark leaderboards. Moreover, we discuss some open challenges and highlight several promising directions, with the hope of inspiring and facilitating more works on these topics for future research.
引用
收藏
页码:178 / 317
页数:140
相关论文
共 50 条
  • [41] QUESTION ANSWERING SYSTEM BASED ON PRE-TRAINING MODEL AND RETRIEVAL RERANKING FOR INDUSTRY 4.0
    Chen, Ta-Fu
    Lin, Yi-Xing
    Su, Ming-Hsiang
    Chen, Po-Kai
    Tai, Tzu-Chiang
    Wang, Jia-Ching
    [J]. 2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 2178 - 2181
  • [42] Unsupervised bin-wise pre-training: A fusion of information theory and hypergraph
    Glory, Anila H.
    Vigneswaran, C.
    Sriram, Shankar V. S.
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 195
  • [43] Heterogeneous graph convolutional network pre-training as side information for improving recommendation
    Do, Phuc
    Pham, Phu
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (18): : 15945 - 15961
  • [44] Heterogeneous graph convolutional network pre-training as side information for improving recommendation
    Phuc Do
    Phu Pham
    [J]. Neural Computing and Applications, 2022, 34 : 15945 - 15961
  • [45] Pre-training neural machine translation with alignment information via optimal transport
    Su, Xueping
    Zhao, Xingkai
    Ren, Jie
    Li, Yunhong
    Raetsch, Matthias
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (16) : 48377 - 48397
  • [46] Pre-training neural machine translation with alignment information via optimal transport
    Xueping Su
    Xingkai Zhao
    Jie Ren
    Yunhong Li
    Matthias Rätsch
    [J]. Multimedia Tools and Applications, 2024, 83 : 48377 - 48397
  • [47] Roles of pre-training in deep neural networks from information theoretical perspective
    Furusho, Yasutaka
    Kubo, Takatomi
    Ikeda, Kazushi
    [J]. NEUROCOMPUTING, 2017, 248 : 76 - 79
  • [48] Improving fault localization with pre-training
    Zhang, Zhuo
    Li, Ya
    Xue, Jianxin
    Mao, Xiaoguang
    [J]. FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (01)
  • [49] Structural Information Guided Multimodal Pre-training for Vehicle-Centric Perception
    Wang, Xiao
    Wu, Wentao
    Li, Chenglong
    Zhao, Zhicheng
    Chen, Zhe
    Shi, Yukai
    Tang, Jin
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 5624 - 5632
  • [50] A Pipelined Pre-training Algorithm for DBNs
    Ma, Zhiqiang
    Li, Tuya
    Yang, Shuangtao
    Zhang, Li
    [J]. CHINESE COMPUTATIONAL LINGUISTICS AND NATURAL LANGUAGE PROCESSING BASED ON NATURALLY ANNOTATED BIG DATA, CCL 2017, 2017, 10565 : 48 - 59