Low-Resource Compositional Semantic Parsing with Concept Pretraining

被引:0
|
作者
Rongali, Subendhu [1 ]
Sridhar, Mukund [2 ]
Khan, Haidar [1 ]
Arkoudas, Konstantine [3 ]
Hamza, Wael [1 ]
McCallum, Andrew [4 ]
机构
[1] Amazon Alexa AI, Sunnyvale, CA 94110 USA
[2] Google, New York, NY USA
[3] Dyania Hlth, Athens, Greece
[4] UMass Amherst, Amherst, MA USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Semantic parsing plays a key role in digital voice assistants such as Alexa, Siri, and Google Assistant by mapping natural language to structured meaning representations. To extend the capabilities of a voice assistant for a new domain, the underlying semantic parsing model needs to be retrained using thousands of annotated examples from the new domain, which is time-consuming and expensive. In this work, we present an architecture to perform such domain adaptation automatically, with only a small amount of metadata about the new domain and without any new training data (zeroshot) or with very few examples (few-shot). We use a base seq2seq (sequence-to-sequence) architecture and augment it with a concept encoder that encodes intent and slot tags from the new domain. We also introduce a novel decoder-focused approach to pretrain seq2seq models to be concept aware using Wikidata. This pretraining helps our model learn important concepts and perform well in low-resource settings. We report few-shot and zero-shot results for compositional semantic parsing on the TOPv2 dataset and show that our model outperforms prior approaches in few-shot settings for the TOPv2 and SNIPS datasets.
引用
收藏
页码:1410 / 1419
页数:10
相关论文
共 50 条
  • [1] Low-Resource Domain Adaptation for Compositional Task-Oriented Semantic Parsing
    Chen, Xilun
    Ghoshal, Asish
    Mehdad, Yashar
    Zettlemoyer, Luke
    Gupta, Sonal
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 5090 - 5100
  • [2] The Power of Prompt Tuning for Low-Resource Semantic Parsing
    Schucher, Nathan
    Reddy, Siva
    de Vries, Harm
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 148 - 156
  • [3] Systematic Investigation of Strategies Tailored for Low-Resource Settings for Low-Resource Dependency Parsing
    Sandhan, Jivnesh
    Behera, Laxmidhar
    Goyal, Pawan
    [J]. 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2164 - 2171
  • [4] A systematic comparison of methods for low-resource dependency parsing on genuinely low-resource languages
    Vania, Clara
    Kementchedjhieva, Yova
    Sogaard, Anders
    Lopez, Adam
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1105 - 1116
  • [5] A Little Pretraining Goes a LongWay: A Case Study on Dependency Parsing Task for Low-resource Morphologically Rich Languages
    Sandhan, Jivnesh
    Krishna, Amrith
    Gupta, Ashim
    Behera, Laxmidhar
    Goyal, Pawan
    [J]. EACL 2021: THE 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2021, : 111 - 120
  • [6] Neural Semantic Parsing in Low-Resource Settings with Back-Translation and Meta-Learning
    Sun, Yibo
    Tang, Duyu
    Duan, Nan
    Gong, Yeyun
    Feng, Xiaocheng
    Qin, Bing
    Jiang, Daxin
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 8960 - 8967
  • [7] Low-Resource Semantic Role Labeling
    Gormley, Matthew R.
    Mitchell, Margaret
    Van Durme, Benjamin
    Dredze, Mark
    [J]. PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2014, : 1177 - 1187
  • [8] Concept Matching for Low-Resource Classification
    Errica, Federico
    Silvestri, Fabrizio
    Edizel, Bora
    Denoyer, Ludovic
    Petroni, Fabio
    Plachouras, Vassilis
    Riedel, Sebastian
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [9] Pretraining by Backtranslation for End-to-end ASR in Low-Resource Settings
    Wiesner, Matthew
    Renduchintala, Adithya
    Watanabe, Shinji
    Liu, Chunxi
    Dehak, Najim
    Khudanpur, Sanjeev
    [J]. INTERSPEECH 2019, 2019, : 4375 - 4379
  • [10] ANALYZING ASR PRETRAINING FOR LOW-RESOURCE SPEECH-TO-TEXT TRANSLATION
    Stoian, Mihaela C.
    Bansal, Sameer
    Goldwater, Sharon
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7909 - 7913