Leveraging Document-Level and Query-Level Passage Cumulative Gain for Document Ranking

被引:0
|
作者
Zhi-Jing Wu
Yi-Qun Liu
Jia-Xin Mao
Min Zhang
Shao-Ping Ma
机构
[1] Tsinghua University,Department of Computer Science and Technology
[2] Tsinghua University,Beijing National Research Center for Information Science and Technology
[3] Renmin University of China,Gaoling School of Artificial Intelligence
关键词
document ranking; neural network; passage cumulative gain;
D O I
暂无
中图分类号
学科分类号
摘要
Document ranking is one of the most studied but challenging problems in information retrieval (IR). More and more studies have begun to address this problem from fine-grained document modeling. However, most of them focus on context-independent passage-level relevance signals and ignore the context information. In this paper, we investigate how information gain accumulates with passages and propose the context-aware Passage Cumulative Gain (PCG). The fine-grained PCG avoids the need to split documents into independent passages. We investigate PCG patterns at the document level (DPCG) and the query level (QPCG). Based on the patterns, we propose a BERT-based sequential model called Passage-level Cumulative Gain Model (PCGM) and show that PCGM can effectively predict PCG sequences. Finally, we apply PCGM to the document ranking task using two approaches. The first one is leveraging DPCG sequences to estimate the gain of an individual document. Experimental results on two public ad hoc retrieval datasets show that PCGM outperforms most existing ranking models. The second one considers the cross-document effects and leverages QPCG sequences to estimate the marginal relevance. Experimental results show that predicted results are highly consistent with users’ preferences. We believe that this work contributes to improving ranking performance and providing more explainability for document ranking.
引用
收藏
页码:814 / 838
页数:24
相关论文
共 50 条
  • [21] GeneTUKit: a software for document-level gene normalization
    Huang, Minlie
    Liu, Jingchen
    Zhu, Xiaoyan
    BIOINFORMATICS, 2011, 27 (07) : 1032 - 1033
  • [22] Discriminative Reasoning for Document-level Relation Extraction
    Xu, Wang
    Chen, Kehai
    Zhao, Tiejun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1653 - 1663
  • [23] Anaphor Assisted Document-Level Relation Extraction
    Lu, Chonggang
    Zhang, Richong
    Sun, Kai
    Kim, Jaein
    Zhang, Cunwang
    Mao, Yongyi
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 15453 - 15464
  • [24] Document-level relation extraction with three channels
    Zhang, Zhanjun
    Zhao, Shan
    Zhang, Haoyu
    Wan, Qian
    Liu, Jie
    KNOWLEDGE-BASED SYSTEMS, 2024, 284
  • [25] Clinical Concept Extraction for Document-Level Coding
    Wiegreffe, Sarah
    Choi, Edward
    Yan, Sherry
    Sun, Jimeng
    Eisenstein, Jacob
    SIGBIOMED WORKSHOP ON BIOMEDICAL NATURAL LANGUAGE PROCESSING (BIONLP 2019), 2019, : 261 - 272
  • [26] A survey on personalized document-level sentiment analysis
    Zhu, Wenhao
    Qiu, Jiayue
    Yu, Ziyue
    Luo, Wuman
    NEUROCOMPUTING, 2024, 609
  • [27] Investigating Contextual Influence in Document-Level Translation
    Nayak, Prashanth
    Haque, Rejwanul
    Kelleher, John D.
    Way, Andy
    INFORMATION, 2022, 13 (05)
  • [28] On Event Individuation for Document-Level Information Extraction
    Gantt, William
    Kriz, Reno
    Chen, Yunmo
    Vashishtha, Siddharth
    White, Aaron Steven
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12938 - 12958
  • [29] Document-Level Event Argument Extraction by Leveraging Redundant Information and Closed Boundary Loss
    Zhou, Hanzhang
    Mao, Kezhi
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3041 - 3052
  • [30] Document-level Relation Extraction with Relation Correlations
    Han, Ridong
    Peng, Tao
    Wang, Benyou
    Liu, Lu
    Tiwari, Prayag
    Wan, Xiang
    NEURAL NETWORKS, 2024, 171 : 14 - 24