Enhancing Text2SQL Generation with Syntactic Information and Multi-task Learning

被引:0
|
作者
Li, Haochen [1 ]
Nuo, Minghua [1 ,2 ,3 ]
机构
[1] Inner Mongolia Univ, Coll Comp Sci, Hohhot 010021, Inner Mongolia, Peoples R China
[2] Natl & Local Joint Engn Res Ctr Intelligent Infor, Hohhot 010021, Inner Mongolia, Peoples R China
[3] Inner Mongolia Key Lab Mongolian Informat Proc Te, Hohhot 010021, Inner Mongolia, Peoples R China
基金
中国国家自然科学基金;
关键词
Text2SQL; Graph neural network; syntactic parsing; multi-task Learning;
D O I
10.1007/978-3-031-44213-1_32
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Text2SQL task aims to convert natural language (NL) questions into SQL queries. Most rule-based traditional syntactic parsers are often ignoring the syntactic information in the question and unseen database schemas, which lead to lower generalization ability. To breakthrough these limitations, we propose a novel model Syn-RGAT that leverages syntactic information of questions. Specifically, our model jointly encodes database schemas and questions by mapping them into a graph, which is then passed through a relation-aware graph attention network (RGAT). For the questions, map it as a syntax information graph and use syntactic augmentation module to learn all the nodes representations of the graph. Then the outputs of RGAT and syntactic augmentation module are integrated. Additionally, to assist the model distinguish between different database schemas, we introduce a graph pruning task and form a multi-task framework which shares the encoder of Text2SQL task. We use a heuristic learning method to combine graph pruning tasks with Text2SQL. In experimental part, Syn-RGAT outperforms all baseline models on the Spider dataset, and we further improve the performance more than 8% with BERT.
引用
收藏
页码:377 / 388
页数:12
相关论文
共 50 条
  • [21] Multi-task Learning with Bidirectional Language Models for Text Classification
    Yang, Qi
    Shang, Lin
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [22] Multi-task learning for historical text normalization: Size matters
    Bollmann, Marcel
    Sogaard, Anders
    Bingel, Joachim
    DEEP LEARNING APPROACHES FOR LOW-RESOURCE NATURAL LANGUAGE PROCESSING (DEEPLO), 2018, : 19 - 24
  • [23] CoTexT: Multi-task Learning with Code-Text Transformer
    Long Phan
    Hieu Tran
    Le, Daniel
    Hieu Nguyen
    Anibal, James
    Peltekian, Alec
    Ye, Yanfang
    NLP4PROG 2021: THE 1ST WORKSHOP ON NATURAL LANGUAGE PROCESSING FOR PROGRAMMING (NLP4PROG 2021), 2021, : 40 - 47
  • [24] Multi-task learning using a hybrid representation for text classification
    Guangquan Lu
    Jiangzhang Gan
    Jian Yin
    Zhiping Luo
    Bo Li
    Xishun Zhao
    Neural Computing and Applications, 2020, 32 : 6467 - 6480
  • [25] Multi-task learning using a hybrid representation for text classification
    Lu, Guangquan
    Gan, Jiangzhang
    Yin, Jian
    Luo, Zhiping
    Li, Bo
    Zhao, Xishun
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (11): : 6467 - 6480
  • [26] Multi-Task Learning for Text-dependent Speaker Verification
    Chen, Nanxin
    Qian, Yanmin
    Yu, Kai
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 185 - 189
  • [27] Fact Aware Multi-task Learning for Text Coherence Modeling
    Abhishek, Tushar
    Rawat, Daksh
    Gupta, Manish
    Varma, Vasudeva
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2022, PT II, 2022, 13281 : 340 - 353
  • [28] Enhancing Explainable Recommendations: Integrating Reason Generation and Rating Prediction through Multi-Task Learning
    Zhu, Xingyu
    Xia, Xiaona
    Wu, Yuheng
    Zhao, Wenxu
    APPLIED SCIENCES-BASEL, 2024, 14 (18):
  • [29] Multi-Task Learning with Language Modeling for Question Generation
    Zhou, Wenjie
    Zhang, Minghua
    Wu, Yunfang
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3394 - 3399
  • [30] Binaural Audio Generation via Multi-task Learning
    Li, Sijia
    Liu, Shiguang
    Manocha, Dinesh
    ACM TRANSACTIONS ON GRAPHICS, 2021, 40 (06):