Gaussian Transformer: A Lightweight Approach for Natural Language Inference

被引:0
|
作者
Guo, Maosheng [1 ]
Zhang, Yu [1 ]
Liu, Ting [1 ]
机构
[1] Harbin Inst Technol, Res Ctr Social Comp & Informat Retrieval, Harbin, Heilongjiang, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Natural Language Inference (NLI) is an active research area, where numerous approaches based on recurrent neural networks (RNNs), convolutional neural networks (CNNs), and self-attention networks (SANS) has been proposed. Although obtaining impressive performance, previous recurrent approaches are hard to train in parallel; convolutional models tend to cost more parameters, while self-attention networks are not good at capturing local dependency of texts. To address this problem, we introduce a Gaussian prior to self-attention mechanism, for better modeling the local structure of sentences. Then we propose an efficient RNN/CNN-free architecture named Gaussian Transformer for NLI, which consists of encoding blocks modeling both local and global dependency, high-order interaction blocks collecting the evidence of multi-step inference, and a lightweight comparison block saving lots of parameters. Experiments show that our model achieves new state-of-the-art performance on both SNLI and MultiNLI benchmarks with significantly fewer parameters and considerably less training time. Besides, evaluation using the Hard NLI datasets demonstrates that our approach is less affected by the undesirable annotation artifacts.
引用
收藏
页码:6489 / 6496
页数:8
相关论文
共 50 条
  • [1] Feature Fusion Transformer Network for Natural Language Inference
    Sun, Lei
    Yan, Hengxin
    [J]. PROCEEDINGS OF 2022 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION (IEEE ICMA 2022), 2022, : 1009 - 1014
  • [2] Evaluating Natural Language Inference Models: A Metamorphic Testing Approach
    Jiang, Mingyue
    Bao, Houzhen
    Tu, Kaiyi
    Zhang, Xiao-Yi
    Ding, Zuohua
    [J]. 2021 IEEE 32ND INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING (ISSRE 2021), 2021, : 220 - 230
  • [3] A semantics-aware approach for multilingual natural language inference
    Phuong Le-Hong
    Erik Cambria
    [J]. Language Resources and Evaluation, 2023, 57 : 611 - 639
  • [4] A semantics-aware approach for multilingual natural language inference
    Le-Hong, Phuong
    Cambria, Erik
    [J]. LANGUAGE RESOURCES AND EVALUATION, 2023, 57 (02) : 611 - 639
  • [5] Probing for Bridging Inference in Transformer Language Models
    Pandit, Onkar
    Hou, Yufang
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 4153 - 4163
  • [6] Communication and inference in natural language
    da Costa, Jorge Campos
    [J]. LETRAS DE HOJE-ESTUDOS E DEBATES EM LINGUISTICA LITERATURA E LINGUA PORTUGUESA, 2005, 40 (01): : 107 - 133
  • [7] Natural Language Inference in Coq
    Chatzikyriakidis, Stergios
    Luo, Zhaohui
    [J]. JOURNAL OF LOGIC LANGUAGE AND INFORMATION, 2014, 23 (04) : 441 - 480
  • [8] Natural Language Inference in Coq
    Stergios Chatzikyriakidis
    Zhaohui Luo
    [J]. Journal of Logic, Language and Information, 2014, 23 : 441 - 480
  • [9] A Transformer-based Approach for Translating Natural Language to Bash Commands
    Fu, Quchen
    Teng, Zhongwei
    White, Jules
    Schmidt, Douglas C.
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 1245 - 1248
  • [10] Dialogue Natural Language Inference
    Welleck, Sean
    Weston, Jason
    Szlam, Arthur
    Cho, Kyunghyun
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3731 - 3741