Explaining pretrained language models' understanding of linguistic structures using construction grammar

被引:0
|
作者
Weissweiler, Leonie [1 ,2 ]
Hofmann, Valentin [1 ,3 ]
Koeksal, Abdullatif [1 ,2 ]
Schuetze, Hinrich [1 ,2 ]
机构
[1] Ludwig Maximilians Univ Munchen, Ctr Informat & Language Proc, Munich, Germany
[2] Munich Ctr Machine Learning, Munich, Germany
[3] Univ Oxford, Fac Linguist, Oxford, England
来源
基金
欧洲研究理事会;
关键词
NLP; probing; construction grammar; computational linguistics; large language models; COMPARATIVE CORRELATIVES;
D O I
10.3389/frai.2023.1225791
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Construction Grammar (CxG) is a paradigm from cognitive linguistics emphasizing the connection between syntax and semantics. Rather than rules that operate on lexical items, it posits constructions as the central building blocks of language, i.e., linguistic units of different granularity that combine syntax and semantics. As a first step toward assessing the compatibility of CxG with the syntactic and semantic knowledge demonstrated by state-of-the-art pretrained language models (PLMs), we present an investigation of their capability to classify and understand one of the most commonly studied constructions, the English comparative correlative (CC). We conduct experiments examining the classification accuracy of a syntactic probe on the one hand and the models' behavior in a semantic application task on the other, with BERT, RoBERTa, and DeBERTa as the example PLMs. Our results show that all three investigated PLMs, as well as OPT, are able to recognize the structure of the CC but fail to use its meaning. While human-like performance of PLMs on many NLP tasks has been alleged, this indicates that PLMs still suffer from substantial shortcomings in central domains of linguistic knowledge.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] ReGen: Reinforcement Learning for Text and Knowledge Base Generation using Pretrained Language Models
    Dognin, Pierre L.
    Padhi, Inkit
    Melnyk, Igor
    Das, Payel
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1084 - 1099
  • [32] Normalized difference vegetation index prediction using reservoir computing and pretrained language models
    Olamofe, John
    Ray, Ram
    Dong, Xishuang
    Qian, Lijun
    ARTIFICIAL INTELLIGENCE IN AGRICULTURE, 2025, 15 (01): : 116 - 129
  • [33] Comparative Study of Multiclass Text Classification in Research Proposals Using Pretrained Language Models
    Lee, Eunchan
    Lee, Changhyeon
    Ahn, Sangtae
    APPLIED SCIENCES-BASEL, 2022, 12 (09):
  • [34] Knowledge Graph Completion for Power Grid Main Equipment Using Pretrained Language Models
    Lin, Chenxiang
    Zheng, Zhou
    Cai, Shitao
    Fu, Li
    Xie, Wei
    Ma, Teng
    Zhang, Zhihong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT IV, 2023, 14089 : 828 - 838
  • [35] Intent-based Product Collections for E-commerce using Pretrained Language Models
    Kim, Hiun
    Jeong, Jisu
    Kim, Kyung-Min
    Lee, Dongjun
    Lee, Hyun Dong
    Seo, Dongpil
    Han, Jeeseung
    Park, Dong Wook
    Heo, Ji Ae
    Kim, Rak Yeong
    21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 228 - 237
  • [36] Automatic Component Prediction for Issue Reports Using Fine-Tuned Pretrained Language Models
    Wang, Dae-Sung
    Lee, Chan-Gun
    IEEE ACCESS, 2022, 10 : 131456 - 131468
  • [37] Identifying Speakers in Dialogue Transcripts: A Text-based Approach Using Pretrained Language Models
    Nguyen, Minh
    Dernoncourt, Franck
    Yoon, Seunghyun
    Deilamsalehy, Hanieh
    Tana, Hao
    Rossi, Ryan
    Trani, Quan Hung
    Bui, Trung
    Nguyen, Thien Huu
    INTERSPEECH 2024, 2024, : 3799 - 3803
  • [38] CancerGPT for few shot drug pair synergy prediction using large pretrained language models
    Li, Tianhao
    Shetty, Sandesh
    Kamath, Advaith
    Jaiswal, Ajay
    Jiang, Xiaoqian
    Ding, Ying
    Kim, Yejin
    NPJ DIGITAL MEDICINE, 2024, 7 (01)
  • [39] PRCBERT: Prompt Learning for Requirement Classification using BERT-based Pretrained Language Models
    Luo, Xianchang
    Xue, Yinxing
    Xing, Zhenchang
    Sun, Jiamou
    PROCEEDINGS OF THE 37TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2022, 2022,
  • [40] CancerGPT for few shot drug pair synergy prediction using large pretrained language models
    Tianhao Li
    Sandesh Shetty
    Advaith Kamath
    Ajay Jaiswal
    Xiaoqian Jiang
    Ying Ding
    Yejin Kim
    npj Digital Medicine, 7