Efficient Open Domain Question Answering With Delayed Attention in Transformer-Based Models

被引:0
|
作者
Siblini, Wissam [1 ]
Challal, Mohamed [1 ]
Pasqual, Charlotte [1 ]
机构
[1] Worldline, Puteaux La Defense, France
关键词
Bert; Deep Learning; Information Retrieval; Knowledge Management; Natural Language Processing; Question Answering; Scalability; Speed; Squad; Transformer;
D O I
10.4018/IJDWM.298005
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Open domain question answering (ODQA) on a large-scale corpus of documents (e.g., Wikipedia) is a key challenge in computer science. Although transformer-based language models such as Bert have shown an ability to outperform humans to extract answers from small pre-selected passages of text, they suffer from their high complexity if the search space is much larger. The most common way to deal with this problem is to add a preliminary information retrieval step to strongly filter the corpus and keep only the relevant passages. In this article, the authors consider a more direct and complementary solution that consists of restricting the attention mechanism in transformer-based models to allow a more efficient management of computations. The resulting variants are competitive with the original models on the extractive task and allow, in the ODQA setting, a significant acceleration of predictions and sometimes even an improvement in the quality of response.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Open-Domain Long-Form Question–Answering Using Transformer-Based Pipeline
    Dash A.
    Awachar M.
    Patel A.
    Rudra B.
    [J]. SN Computer Science, 4 (5)
  • [2] A Transformer-based Medical Visual Question Answering Model
    Liu, Lei
    Su, Xiangdong
    Guo, Hui
    Zhu, Daobin
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 1712 - 1718
  • [3] A Survey for Efficient Open Domain Question Answering
    Zhang, Qin
    Chen, Shangsi
    Xu, Dongkuan
    Cao, Qingqing
    Chen, Xiaojun
    Cohn, Trevor
    Fang, Meng
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14447 - 14465
  • [4] A lightweight Transformer-based visual question answering network with Weight-Sharing Hybrid Attention
    Zhu, Yue
    Chen, Dongyue
    Jia, Tong
    Deng, Shizhuo
    [J]. NEUROCOMPUTING, 2024, 608
  • [5] Transformer-Based Extractive Social Media Question Answering on TweetQA
    Butt, Sabur
    Ashraf, Noman
    Fahim, Hammad
    Sidorov, Grigori
    Gelbukh, Alexander
    [J]. COMPUTACION Y SISTEMAS, 2021, 25 (01): : 23 - 32
  • [6] Transformer-Based Neural Network for Answer Selection in Question Answering
    Shao, Taihua
    Guo, Yupu
    Chen, Honghui
    Hao, Zepeng
    [J]. IEEE ACCESS, 2019, 7 : 26146 - 26156
  • [7] SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval
    Zhao, Tiancheng
    Lu, Xiaopeng
    Lee, Kyusong
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 565 - 575
  • [8] Entity-aware answer sentence selection for question answering with transformer-based language models
    Abbasiantaeb, Zahra
    Momtazi, Saeedeh
    [J]. JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2022, 59 (03) : 755 - 777
  • [9] Entity-aware answer sentence selection for question answering with transformer-based language models
    Abbasiantaeb, Zahra
    Momtazi, Saeedeh
    [J]. Journal of Intelligent Information Systems, 2022, 59 (03): : 755 - 777
  • [10] Entity-aware answer sentence selection for question answering with transformer-based language models
    Zahra Abbasiantaeb
    Saeedeh Momtazi
    [J]. Journal of Intelligent Information Systems, 2022, 59 : 755 - 777