Long Document Ranking with Query-Directed Sparse Transformer

被引:0
|
作者
Jiang, Jyun-Yu [1 ]
Xiong, Chenyan [2 ]
Lee, Chia-Jung [3 ]
Wang, Wei [1 ]
机构
[1] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90024 USA
[2] Microsoft Res AI, Redmond, WA USA
[3] Amazon, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The computing cost of transformer selfattention often necessitates breaking long documents to fit in pretrained models in document ranking tasks. In this paper, we design Query-Directed Sparse attention that induces IR-axiomatic structures in transformer self-attention. Our model, QDS-Transformer, enforces the principle properties desired in ranking: local contextualization, hierarchical representation, and query-oriented proximity matching, while it also enjoys efficiency from sparsity. Experiments on one fully supervised and three few-shot TREC document ranking benchmarks demonstrate the consistent and robust advantage of QDSTransformer over previous approaches, as they either retrofit long documents into BERT or use sparse attention without emphasizing IR principles. We further quantify the computing complexity and demonstrates that our sparse attention with TVM implementation is twice more efficient that the fully-connected selfattention. All source codes, trained model, and predictions of this work are available at https://github.com/hallogameboy/ QDS-Transformer.
引用
下载
收藏
页码:4594 / 4605
页数:12
相关论文
共 50 条
  • [31] Query-focused multi-document summarization using hypergraph-based ranking
    Xiong, Shufeng
    Ji, Donghong
    INFORMATION PROCESSING & MANAGEMENT, 2016, 52 (04) : 670 - 681
  • [32] Co-HITS-Ranking Based Query-Focused Multi-document Summarization
    Hu, Po
    Ji, Donghong
    Teng, Chong
    INFORMATION RETRIEVAL TECHNOLOGY, 2010, 6458 : 121 - 130
  • [33] Hi-Transformer: Hierarchical Interactive Transformer for Efficient and Effective Long Document Modeling
    Wu, Chuhan
    Wu, Fangzhao
    Qi, Tao
    Huang, Yongfeng
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 848 - 853
  • [34] GLASSOFORMER: A QUERY-SPARSE TRANSFORMER FOR POST-FAULT POWER GRID VOLTAGE PREDICTION
    Zheng, Yunling
    Hu, Carson
    Lin, Guang
    Yue, Meng
    Wang, Bao
    Xin, Jack
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3968 - 3972
  • [35] Sparse Sampling Transformer with Uncertainty-Driven Ranking for Unified Removal of Raindrops and Rain Streaks
    Chen, Sixiang
    Ye, Tian
    Bai, Jinbin
    Chen, Erkang
    Shi, Jun
    Zhu, Lei
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 13060 - 13071
  • [36] Leveraging Salience Analysis and Sparse Attention for Long Document Summarization
    Jiang, Zhihua
    Chen, Yaxuan
    Rao, Dongning
    PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND INFORMATION RETRIEVAL, NLPIR 2023, 2023, : 44 - 50
  • [37] Long Document Re-ranking with Modular Re-ranker
    Gao, Luyu
    Callan, Jamie
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2371 - 2376
  • [38] Efficient Document-at-a-time and Score-at-a-time Query Evaluation for Learned Sparse Representations
    MacKenzie, Joel
    Trotman, Andrew
    Lin, Jimmy
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2023, 41 (04)
  • [39] Double-Hypergraph based Sentence Ranking for Query-Focused Multi-Document Summarizaton
    Cai, Xiaoyan
    Han, Junwei
    Guo, Lei
    Yang, Libin
    2016 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE WORKSHOPS (WIW 2016), 2016, : 112 - 118
  • [40] A Context-Sensitive Manifold Ranking Approach to Query-Focused Multi-document Summarization
    Cai, Xiaoyan
    Li, Wenjie
    PRICAI 2010: TRENDS IN ARTIFICIAL INTELLIGENCE, 2010, 6230 : 27 - 38