Long Document Ranking with Query-Directed Sparse Transformer

被引:0
|
作者
Jiang, Jyun-Yu [1 ]
Xiong, Chenyan [2 ]
Lee, Chia-Jung [3 ]
Wang, Wei [1 ]
机构
[1] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90024 USA
[2] Microsoft Res AI, Redmond, WA USA
[3] Amazon, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The computing cost of transformer selfattention often necessitates breaking long documents to fit in pretrained models in document ranking tasks. In this paper, we design Query-Directed Sparse attention that induces IR-axiomatic structures in transformer self-attention. Our model, QDS-Transformer, enforces the principle properties desired in ranking: local contextualization, hierarchical representation, and query-oriented proximity matching, while it also enjoys efficiency from sparsity. Experiments on one fully supervised and three few-shot TREC document ranking benchmarks demonstrate the consistent and robust advantage of QDSTransformer over previous approaches, as they either retrofit long documents into BERT or use sparse attention without emphasizing IR principles. We further quantify the computing complexity and demonstrates that our sparse attention with TVM implementation is twice more efficient that the fully-connected selfattention. All source codes, trained model, and predictions of this work are available at https://github.com/hallogameboy/ QDS-Transformer.
引用
收藏
页码:4594 / 4605
页数:12
相关论文
共 50 条
  • [41] HM-Transformer: Hierarchical Multi-modal Transformer for Long Document Image Understanding
    Deng, Xi
    Li, Shasha
    Yu, Jie
    Ma, Jun
    WEB AND BIG DATA, PT IV, APWEB-WAIM 2023, 2024, 14334 : 232 - 245
  • [42] Globalizing BERT-based Transformer Architectures for Long Document Summarization
    Grail, Quentin
    Perez, Julien
    Gaussier, Eric
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1792 - 1810
  • [43] A Query Specific Graph Based Approach to Multi-document Text Summarization: Simultaneous Cluster and Sentence Ranking
    Pandit, Sandip R.
    Potey, M. A.
    2013 INTERNATIONAL CONFERENCE ON MACHINE INTELLIGENCE AND RESEARCH ADVANCEMENT (ICMIRA 2013), 2013, : 213 - 217
  • [44] ERNIE-DOC: A Retrospective Long-Document Modeling Transformer
    Ding, Siyu
    Shang, Junyuan
    Wang, Shuohuan
    Sun, Yu
    Tian, Hao
    Wu, Hua
    Wang, Haifeng
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 2914 - 2927
  • [45] CWRCzech: 100M Query-Document Czech Click Dataset and Its Application toWeb Relevance Ranking
    Vonasek, Josef
    Straka, Milan
    Krc, Rostislav
    Lasonova, Lenka
    Egorova, Ekaterina
    Strakova, Jana
    Naplava, Jakub
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 1221 - 1231
  • [46] Event-Specific Document Ranking Through Multi-stage Query Expansion Using an Event Knowledge Graph
    Abdollahi, Sara
    Kuculo, Tin
    Gottschalk, Simon
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT II, 2024, 14609 : 333 - 348
  • [47] HETFORMER: Heterogeneous Transformer with Sparse Attention for Long-Text Extractive Summarization
    Liu, Ye
    Zhang, Jian-Guo
    Wan, Yao
    Xia, Congying
    He, Lifang
    Yu, Philip S.
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 146 - 154
  • [48] LNLF-BERT: Transformer for Long Document Classification With Multiple Attention Levels
    Pham, Linh Manh
    Cao The, Hoang
    IEEE Access, 2024, 12 : 165348 - 165358
  • [49] Hybrid Attention-based Transformer for Long-range Document Classification
    Qin, Ruyu
    Huang, Min
    Liu, Jiawei
    Miao, Qinghai
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [50] Mutually Reinforced Manifold-Ranking Based Relevance Propagation Model for Query-Focused Multi-Document Summarization
    Cai, Xiaoyan
    Li, Wenjie
    IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2012, 20 (05): : 1597 - 1607