Pre-trained Language Model-based Retrieval and Ranking forWeb Search

被引:4
|
作者
Zou, Lixin [1 ]
Lu, Weixue [1 ]
Liu, Yiding [1 ]
Cai, Hengyi [1 ]
Chu, Xiaokai [1 ]
Ma, Dehong [1 ]
Shi, Daiting [1 ]
Sun, Yu [1 ]
Cheng, Zhicong [1 ]
Gu, Simiu [1 ]
Wang, Shuaiqiang [1 ]
Yin, Dawei [1 ]
机构
[1] Baidu Inc, Beijing, Peoples R China
关键词
Pre-trained language model; web retrieval; ranking; ALGORITHM;
D O I
10.1145/3568681
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Pre-trained language representation models (PLMs) such as BERT and Enhanced Representation through kNowledge IntEgration (ERNIE) have been integral to achieving recent improvements on various downstream tasks, including information retrieval. However, it is nontrivial to directly utilize these models for the large-scale web search due to the following challenging issues: (1) the prohibitively expensive computations ofmassive neural PLMs, especially for long texts in the web document, prohibit their deployments in the web search system that demands extremely low latency; (2) the discrepancy between existing task-agnostic pre-training objectives and the ad hoc retrieval scenarios that demand comprehensive relevance modeling is another main barrier for improving the online retrieval and ranking effectiveness; and (3) to create a significant impact on real-world applications, it also calls for practical solutions to seamlessly interweave the resultant PLM and other components into a cooperative system to serve web-scale data. Accordingly, we contribute a series of successfully applied techniques in tackling these exposed issues in this work when deploying the state-of-the-art Chinese pre-trained language model, i.e., ERNIE, in the online search engine system. We first present novel practices to perform expressive PLM-based semantic retrieval with a flexible poly-interaction scheme and cost-efficiently contextualize and rank web documents with a cheap yet powerful Pyramid-ERNIE architecture. We then endow innovative pre-training and fine-tuning paradigms to explicitly incentivize the query-document relevance modeling in PLM-based retrieval and ranking with the large-scale noisy and biased post-click behavioral data. We also introduce a series of effective strategies to seamlessly interwoven the designed PLM-based models with other conventional components into a cooperative system. Extensive offline and online experimental results show that our proposed techniques are crucial to achieving more effective search performance. We also provide a thorough analysis of our methodology and experimental results.
引用
收藏
页数:36
相关论文
共 50 条
  • [21] ViDeBERTa: A powerful pre-trained language model for Vietnamese
    Tran, Cong Dao
    Pham, Nhut Huy
    Nguyen, Anh
    Hy, Truong Son
    Vu, Tu
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1071 - 1078
  • [22] Misspelling Correction with Pre-trained Contextual Language Model
    Hu, Yifei
    Ting, Xiaonan
    Ko, Youlim
    Rayz, Julia Taylor
    PROCEEDINGS OF 2020 IEEE 19TH INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS & COGNITIVE COMPUTING (ICCI*CC 2020), 2020, : 144 - 149
  • [23] CLEAR: Cross-Transformers With Pre-Trained Language Model for Person Attribute Recognition and Retrieval
    Bui, Doanh C.
    Le, Thinh, V
    Ngo, Ba Hung
    Choi, Tae Jong
    PATTERN RECOGNITION, 2025, 164
  • [24] Leveraging Pre-trained Large Language Models to Construct and Utilize World Models for Model-based Task Planning
    Guan, Lin
    Valmeekam, Karthik
    Sreedharan, Sarath
    Kambhampati, Subbarao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [25] CLIP-Llama: A New Approach for Scene Text Recognition with a Pre-Trained Vision-Language Model and a Pre-Trained Language Model
    Zhao, Xiaoqing
    Xu, Miaomiao
    Silamu, Wushour
    Li, Yanbing
    SENSORS, 2024, 24 (22)
  • [26] DynamicRetriever: A Pre-trained Model-based IR System Without an Explicit Index
    Zhou, Yu-Jia
    Yao, Jing
    Dou, Zhi-Cheng
    Wu, Ledell
    Wen, Ji-Rong
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (02) : 276 - 288
  • [27] Pre-Trained Model-Based NFR Classification: Overcoming Limited Data Challenges
    Rahman, Kiramat
    Ghani, Anwar
    Alzahrani, Abdulrahman
    Tariq, Muhammad Usman
    Rahman, Arif Ur
    IEEE ACCESS, 2023, 11 : 81787 - 81802
  • [28] A text restoration model for ancient texts based on pre-trained language model RoBERTa
    Gu, Zhongyu
    Guan, Yanzhi
    Zhang, Shuai
    PROCEEDINGS OF 2024 4TH INTERNATIONAL CONFERENCE ON INTERNET OF THINGS AND MACHINE LEARNING, IOTML 2024, 2024, : 96 - 102
  • [29] A Pre-trained Language Model for Medical Question Answering Based on Domain Adaption
    Liu, Lang
    Ren, Junxiang
    Wu, Yuejiao
    Song, Ruilin
    Cheng, Zhen
    Wang, Sibo
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT II, 2022, 13552 : 216 - 227
  • [30] A Pre-Trained Language Model Based on LED for Tibetan Long Text Summarization
    Ouyang, Xinpeng
    Yan, Xiaodong
    Hao, Minghui
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 992 - 997