Enhancing SPARQL Query Generation for Knowledge Base Question Answering Systems by Learning to Correct Triplets

被引:0
|
作者
Qi, Jiexing [1 ]
Su, Chang [1 ]
Guo, Zhixin [1 ]
Wu, Lyuwen [1 ]
Shen, Zanwei [1 ]
Fu, Luoyi [1 ]
Wang, Xinbing [1 ]
Zhou, Chenghu [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai 200240, Peoples R China
[2] Chinese Acad Sci, Inst Geog Sci & Nat Resources Res, Beijing 100101, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 04期
关键词
Knowledge Base Question Answering; Text-to-SPARQL; semantic parsing; further pretraining; Triplet Structure;
D O I
10.3390/app14041521
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Generating SPARQL queries from natural language questions is challenging in Knowledge Base Question Answering (KBQA) systems. The current state-of-the-art models heavily rely on fine-tuning pretrained models such as T5. However, these methods still encounter critical issues such as triple-flip errors (e.g., (subject, relation, object) is predicted as (object, relation, subject)). To address this limitation, we introduce TSET (Triplet Structure Enhanced T5), a model with a novel pretraining stage positioned between the initial T5 pretraining and the fine-tuning for the Text-to-SPARQL task. In this intermediary stage, we introduce a new objective called Triplet Structure Correction (TSC) to train the model on a SPARQL corpus derived from Wikidata. This objective aims to deepen the model's understanding of the order of triplets. After this specialized pretraining, the model undergoes fine-tuning for SPARQL query generation, augmenting its query-generation capabilities. We also propose a method named "semantic transformation" to fortify the model's grasp of SPARQL syntax and semantics without compromising the pre-trained weights of T5. Experimental results demonstrate that our proposed TSET outperforms existing methods on three well-established KBQA datasets: LC-QuAD 2.0, QALD-9 plus, and QALD-10, establishing a new state-of-the-art performance (95.0% F1 and 93.1% QM on LC-QuAD 2.0, 75.85% F1 and 61.76% QM on QALD-9 plus, 51.37% F1 and 40.05% QM on QALD-10).
引用
收藏
页数:19
相关论文
共 50 条
  • [41] Knowledge Base Question Answering With Attentive Pooling for Question Representation
    Wang, Run-Ze
    Ling, Zhen-Hua
    Hu, Yu
    IEEE ACCESS, 2019, 7 : 46773 - 46784
  • [42] Query Path Generation via Bidirectional Reasoning for Multihop Question Answering From Knowledge Bases
    Zhang, Geng
    Liu, Jin
    Zhou, Guangyou
    Xie, Zhiwen
    Yu, Xiao
    Cui, Xiaohui
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (03) : 1183 - 1195
  • [43] Improving complex knowledge base question answering via structural information learning
    Zhang, Jinhao
    Zhang, Lizong
    Hui, Bei
    Tian, Ling
    KNOWLEDGE-BASED SYSTEMS, 2022, 242
  • [44] Few-shot In-context Learning for Knowledge Base Question Answering
    Li, Tianle
    Ma, Xueguang
    Zhuang, Alex
    Gu, Yu
    Su, Yu
    Chen, Wenhu
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 6966 - 6980
  • [45] Straight to the Facts: Learning Knowledge Base Retrieval for Factual Visual Question Answering
    Narasimhan, Medhini
    Schwing, Alexander G.
    COMPUTER VISION - ECCV 2018, PT VIII, 2018, 11212 : 460 - 477
  • [46] Complex Query Augmentation for Question Answering over Knowledge Graphs
    Abdelkawi, Abdelrahman
    Zafar, Hamid
    Maleshkova, Maria
    Lehmann, Jens
    ON THE MOVE TO MEANINGFUL INTERNET SYSTEMS: OTM 2019 CONFERENCES, 2019, 11877 : 571 - 587
  • [47] Deep Query Ranking for Question Answering over Knowledge Bases
    Zafar, Hamid
    Napolitano, Giulio
    Lehmann, Jens
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2018, PT III, 2019, 11053 : 635 - 638
  • [48] RNG-KBQA: Generation Augmented Iterative Ranking for Knowledge Base Question Answering
    Ye, Xi
    Yavuz, Semih
    Hashimoto, Kazuma
    Zhou, Yingbo
    Xiong, Caiming
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 6032 - 6043
  • [49] Knowledge Base Question Answering through Recursive Hypergraphs
    Yadati, Naganand
    Dayanidhi, R.
    Vaishnavi, S.
    Indira, S.
    Srinidhi, S.
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 448 - 454
  • [50] Question Answering System based on Diease Knowledge Base
    Wang, Xuan
    Wang, Zhijun
    PROCEEDINGS OF 2020 IEEE 11TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND SERVICE SCIENCE (ICSESS 2020), 2020, : 351 - 354