Long Document Ranking with Query-Directed Sparse Transformer

被引:0
|
作者
Jiang, Jyun-Yu [1 ]
Xiong, Chenyan [2 ]
Lee, Chia-Jung [3 ]
Wang, Wei [1 ]
机构
[1] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90024 USA
[2] Microsoft Res AI, Redmond, WA USA
[3] Amazon, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The computing cost of transformer selfattention often necessitates breaking long documents to fit in pretrained models in document ranking tasks. In this paper, we design Query-Directed Sparse attention that induces IR-axiomatic structures in transformer self-attention. Our model, QDS-Transformer, enforces the principle properties desired in ranking: local contextualization, hierarchical representation, and query-oriented proximity matching, while it also enjoys efficiency from sparsity. Experiments on one fully supervised and three few-shot TREC document ranking benchmarks demonstrate the consistent and robust advantage of QDSTransformer over previous approaches, as they either retrofit long documents into BERT or use sparse attention without emphasizing IR principles. We further quantify the computing complexity and demonstrates that our sparse attention with TVM implementation is twice more efficient that the fully-connected selfattention. All source codes, trained model, and predictions of this work are available at https://github.com/hallogameboy/ QDS-Transformer.
引用
收藏
页码:4594 / 4605
页数:12
相关论文
共 50 条
  • [21] A Neural Autoencoder Approach for Document Ranking and Query Refinement in Pharmacogenomic Information Retrieval
    Pfeiffer, Jonas
    Broscheit, Samuel
    Gemulla, Rainer
    Goeschl, Mathias
    SIGBIOMED WORKSHOP ON BIOMEDICAL NATURAL LANGUAGE PROCESSING (BIONLP 2018), 2018, : 87 - 97
  • [22] Query-driven Segment Selection for Ranking Long Documents
    Kim, Youngwoo
    Rahimi, Razieh
    Bonab, Hamed
    Allan, James
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3147 - 3151
  • [23] Improving Transformer-Kernel Ranking Model Using Conformer and Query Term Independence
    Mitra, Bhaskar
    Hofstatter, Sebastian
    Zamani, Hamed
    Craswell, Nick
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 1697 - 1702
  • [24] Generating Relevant and Diverse Query Suggestions Using Sparse Manifold Ranking with Sink Regions
    Van Thanh Nguyen
    Kim Anh Nguyen
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS (ACIIDS 2013), PT I,, 2013, 7802 : 176 - 185
  • [25] The Opposite of Smoothing: A Language Model Approach to Ranking Query-Specific Document Clusters
    Kurland, Oren
    Krikon, Eyal
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2011, 41 : 367 - 395
  • [26] A ranking algorithm for query expansion based on the term's appearing probability in the single document
    Chou, Shihchieh
    Cheng, Chinyi
    Huang, Szujui
    ONLINE INFORMATION REVIEW, 2011, 35 (02) : 217 - 236
  • [27] DREQ: Document Re-ranking Using Entity-Based Query Understanding
    Chatterjee, Shubham
    Mackie, Iain
    Dalton, Jeff
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT I, 2024, 14608 : 210 - 229
  • [28] Sparse Transformer Hawkes Process for Long Event Sequences
    Li, Zhuoqun
    Sun, Mingxuan
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT V, 2023, 14173 : 172 - 188
  • [29] Hierarchical Attention Transformer Networks for Long Document Classification
    Hu, Yongli
    Chen, Puman
    Liu, Tengfei
    Gao, Junbin
    Sun, Yanfeng
    Yin, Baocai
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [30] A Query-Sensitive Graph-Based Sentence Ranking Algorithm for Query-Oriented Multi-Document Summarization .
    Wei, Furu
    He, Yanxiang
    Li, Wenjie
    Lu, Qin
    2008 INTERNATIONAL SYMPOSIUM ON INFORMATION PROCESSING AND 2008 INTERNATIONAL PACIFIC WORKSHOP ON WEB MINING AND WEB-BASED APPLICATION, 2008, : 9 - +