Enhancing yes/no question answering with weak supervision via extractive question answering

被引:1
|
作者
Dimitriadis, Dimitris [1 ]
Tsoumakas, Grigorios [1 ]
机构
[1] Aristotle Univ Thessaloniki, Sch Informat, Thessaloniki 54124, Greece
关键词
Question answering; Yes/no question answering; Extractive question answering; Transformers;
D O I
10.1007/s10489-023-04751-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The effectiveness of natural language processing models relies on various factors, including the architecture, number of parameters, data used during training, and the tasks they were trained on. Recent studies indicate that models pre-trained on large corpora and fine-tuned on task-specific datasets, covering multiple tasks, can generate remarkable results across various benchmarks. We propose a new approach based on a straightforward hypothesis: improving model performance on a target task by considering other artificial tasks defined on the same training dataset. By doing so, the model can gain further insights into the training dataset and attain a greater understanding, improving efficiency on the target task. This approach differs from others that consider multiple pre-existing tasks on different datasets. We validate this hypothesis by focusing on the problem of answering yes/no questions and introducing a multi-task model that outputs a span of the reference text, serving as evidence for answering the question. The task of span extraction is an artificial one, designed to benefit the performance of the model answering yes/no questions. We acquire weak supervision for these spans, by using a pre-trained extractive question answering model, dispensing the need for costly human annotation. Our experiments, using modern transformer-based language models, demonstrate that this method outperforms the standard approach of training models to answer yes/no questions. Although the primary objective was to enhance the performance of the model in answering yes/no questions, it was discovered that span texts are a significant source of information. These spans, derived from the question reference texts, provided valuable insights for the users to better comprehend the answers to the questions. The model's improved accuracy in answering yes/no questions, coupled with the supplementary information provided by the span texts, led to a more comprehensive and informative user experience.
引用
收藏
页码:27560 / 27570
页数:11
相关论文
共 50 条
  • [1] Enhancing yes/no question answering with weak supervision via extractive question answering
    Dimitris Dimitriadis
    Grigorios Tsoumakas
    [J]. Applied Intelligence, 2023, 53 : 27560 - 27570
  • [2] WeaQA: Weak Supervision via Captions for Visual Question Answering
    Banerjee, Pratyay
    Gokhale, Tejas
    Yang, Yezhou
    Baral, Chitta
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3420 - 3435
  • [3] INTERPRETABLE VISUAL QUESTION ANSWERING VIA REASONING SUPERVISION
    Parelli, Maria
    Mallis, Dimitrios
    Diomataris, Markos
    Pitsikalis, Vassilis
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 2525 - 2529
  • [4] Yes/No Question Answering in BioASQ 2019
    Dimitriadis, Dimitris
    Tsoumakas, Grigorios
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 1168 : 661 - 669
  • [5] Factoid Question Answering with Distant Supervision
    Zhang, Hongzhi
    Liang, Xiao
    Xu, Guangluan
    Fu, Kun
    Li, Feng
    Huang, Tinglei
    [J]. ENTROPY, 2018, 20 (06)
  • [6] Sequence tagging for biomedical extractive question answering
    Yoon, Wonjin
    Jackson, Richard
    Lagerberg, Aron
    Kang, Jaewoo
    [J]. BIOINFORMATICS, 2022, 38 (15) : 3794 - 3801
  • [7] QUASER: Question Answering with Scalable Extractive Rationalization
    Ghoshal, Asish
    Iyer, Srinivasan
    Paranjape, Bhargavi
    Lakhotia, Kushal
    Yih, Scott Wen-tau
    Mehdad, Yashar
    [J]. PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1208 - 1218
  • [9] Neural Ranking with Weak Supervision for Open-Domain Question Answering : A Survey
    Shen, Xiaoyu
    Vakulenko, Svitlana
    del Tredici, Marco
    Barlacchi, Gianni
    Byrne, Bill
    de Gispert, Adria
    [J]. 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1736 - 1750
  • [10] Guiding Extractive Summarization with Question-Answering Rewards
    Arumae, Kristjan
    Liu, Fei
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2566 - 2577