Long Document Re-ranking with Modular Re-ranker

被引:2
|
作者
Gao, Luyu [1 ]
Callan, Jamie [1 ]
机构
[1] Carnegie Mellon Univ, Language Technol Inst, Pittsburgh, PA 15213 USA
基金
美国国家科学基金会;
关键词
Neural IR; Document Re-ranking; Deep Learning;
D O I
10.1145/3477495.3531860
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Long document re-ranking has been a challenging problem for neural re-rankers based on deep language models like BERT. Early work breaks the documents into short passage-like chunks. These chunks are independently mapped to scalar scores or latent vectors, which are then pooled into a final relevance score. These encode-and-pool methods however inevitably introduce an information bottleneck: the low dimension representations. In this paper, we propose instead to model full query-to-document interaction, lever-aging the attention operation and modular Transformer re-ranker framework. First, document chunks are encoded independently with an encoder module. An interaction module then encodes the query and performs joint attention from the query to all document chunk representations. We demonstrate that the model can use this newdegree of freedom to aggregate important information from the entire document. Our experiments show that this design produces effective re-ranking on two classical IR collections Robust04 and ClueWeb09, and a large-scale supervised collection MS-MARCO document ranking.1
引用
下载
收藏
页码:2371 / 2376
页数:6
相关论文
共 50 条
  • [1] Re-ranking model based on document clusters
    Lee, KS
    Park, YC
    Choi, KS
    INFORMATION PROCESSING & MANAGEMENT, 2001, 37 (01) : 1 - 14
  • [2] Document Re-ranking via the EM Algorithm
    Teng, Chong
    He, Yanxiang
    Ji, Donghong
    Han, Yue
    RECENT ADVANCES OF ASIAN LANGUAGE PROCESSING TECHNOLOGIES, 2008, : 146 - +
  • [3] Semantic Relatedness Based Re-ranker for Text Spotting
    Sabir, Ahmed
    Moreno-Noguer, Francesc
    Padro, Lluis
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3451 - 3457
  • [4] A Study on Pseudo Labeled Document Constructed for Document Re-ranking
    Teng, Chong
    He, Yanxiang
    Ji, Donghong
    Lin, Guimin
    Mai, Zhewei
    2009 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND COMPUTATIONAL INTELLIGENCE, VOL III, PROCEEDINGS, 2009, : 377 - +
  • [5] A Syntax-Aware Re-ranker for Microblog Retrieval
    Severyn, Aliaksei
    Moschitti, Alessandro
    Tsagkias, Manos
    Berendsen, Richard
    de Rijke, Maarten
    SIGIR'14: PROCEEDINGS OF THE 37TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2014, : 1067 - 1070
  • [6] Document Re-ranking using Partial Social Tagging
    Li, Peng
    Nie, Jian-Yun
    Wang, Bin
    He, Jing
    2012 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT 2012), VOL 1, 2012, : 274 - 281
  • [7] A RE-RANKER SCHEME FOR INTEGRATING LARGE SCALE NLU MODELS
    Su, Chengwei
    Gupta, Rahul
    Ananthakrishnan, Shankar
    Matsoukas, Spyros
    2018 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2018), 2018, : 670 - 676
  • [8] Regression by Re-Ranking
    Goncalves, Filipe Marcel Fernandes
    Pedronette, Daniel Carlos Guimaraes
    Torres, Ricardo da Silva
    PATTERN RECOGNITION, 2023, 140
  • [9] Re-ranking method based on inter-document distances
    Balinski, J
    Danilowicz, C
    INFORMATION PROCESSING & MANAGEMENT, 2005, 41 (04) : 759 - 775
  • [10] Entropy-based Clustering for Improving Document Re-ranking
    Teng, Chong
    He, Yanxiang
    Ji, Donghong
    zhou, Cheng
    Geng, Yixuan
    Chen, Shu
    2009 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND INTELLIGENT SYSTEMS, PROCEEDINGS, VOL 3, 2009, : 662 - +