Explaining pretrained language models' understanding of linguistic structures using construction grammar

被引:0
|
作者
Weissweiler, Leonie [1 ,2 ]
Hofmann, Valentin [1 ,3 ]
Koeksal, Abdullatif [1 ,2 ]
Schuetze, Hinrich [1 ,2 ]
机构
[1] Ludwig Maximilians Univ Munchen, Ctr Informat & Language Proc, Munich, Germany
[2] Munich Ctr Machine Learning, Munich, Germany
[3] Univ Oxford, Fac Linguist, Oxford, England
来源
基金
欧洲研究理事会;
关键词
NLP; probing; construction grammar; computational linguistics; large language models; COMPARATIVE CORRELATIVES;
D O I
10.3389/frai.2023.1225791
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Construction Grammar (CxG) is a paradigm from cognitive linguistics emphasizing the connection between syntax and semantics. Rather than rules that operate on lexical items, it posits constructions as the central building blocks of language, i.e., linguistic units of different granularity that combine syntax and semantics. As a first step toward assessing the compatibility of CxG with the syntactic and semantic knowledge demonstrated by state-of-the-art pretrained language models (PLMs), we present an investigation of their capability to classify and understand one of the most commonly studied constructions, the English comparative correlative (CC). We conduct experiments examining the classification accuracy of a syntactic probe on the one hand and the models' behavior in a semantic application task on the other, with BERT, RoBERTa, and DeBERTa as the example PLMs. Our results show that all three investigated PLMs, as well as OPT, are able to recognize the structure of the CC but fail to use its meaning. While human-like performance of PLMs on many NLP tasks has been alleged, this indicates that PLMs still suffer from substantial shortcomings in central domains of linguistic knowledge.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Using Natural Sentences for Understanding Biases in Language Models
    Alnegheimish, Sarah
    Guo, Alicia
    Sun, Yi
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 2824 - 2830
  • [42] Language understanding using n-multigram models
    Hurtado, L
    Segarra, E
    García, F
    Sanchis, E
    ADVANCES IN NATURAL LANGUAGE PROCESSING, 2004, 3230 : 207 - 219
  • [43] Alzheimer's Disease Detection from Spontaneous Speech through Combining Linguistic Complexity and (Dis)Fluency Features with Pretrained Language Models
    Qiao, Yu
    Yin, Xuefeng
    Wiechmann, Daniel
    Kerz, Elma
    INTERSPEECH 2021, 2021, : 3805 - 3809
  • [44] Refactoring goal-oriented models: a linguistic improvement using large language models
    Alturayeif, Nouf
    Hassine, Jameleddine
    SOFTWARE AND SYSTEMS MODELING, 2025,
  • [45] Automatic uncovering of patient primary concerns in portal messages using a fusion framework of pretrained language models
    Ren, Yang
    Wu, Yuqi
    Fan, Jungwei W.
    Khurana, Aditya
    Fu, Sunyang
    Wu, Dezhi
    Liu, Hongfang
    Huang, Ming
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (08) : 1714 - 1724
  • [46] Explaining machine learning models with interactive natural language conversations using TalkToModel
    Slack, Dylan
    Krishna, Satyapriya
    Lakkaraju, Himabindu
    Singh, Sameer
    NATURE MACHINE INTELLIGENCE, 2023, 5 (08) : 873 - +
  • [47] Explaining machine learning models with interactive natural language conversations using TalkToModel
    Dylan Slack
    Satyapriya Krishna
    Himabindu Lakkaraju
    Sameer Singh
    Nature Machine Intelligence, 2023, 5 : 873 - 883
  • [48] AmericasNLI: Evaluating Zero-shot Natural Language Understanding of Pretrained Multilingual Models in Truly Low-resource Languages
    Ebrahimi, Abteen
    Mager, Manuel
    Oncevay, Arturo
    Chaudhary, Vishrav
    Chiruzzo, Luis
    Fan, Angela
    Ortega, John E.
    Ramos, Ricardo
    Rios, Annette
    Meza-Ruiz, Ivan
    Gimenez-Lugo, Gustavo A.
    Mager, Elisabeth
    Neubig, Graham
    Palmer, Alexis
    Coto-Solano, Rolando
    Ngoc Thang Vu
    Kann, Katharina
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 6279 - 6299
  • [49] CuPe-KG: Cultural perspective-based knowledge graph construction of tourism resources via pretrained language models
    Fan, Zhanling
    Chen, Chongcheng
    INFORMATION PROCESSING & MANAGEMENT, 2024, 61 (03)
  • [50] Identifying B-cell epitopes using AlphaFold2 predicted structures and pretrained language model
    Zeng, Yuansong
    Wei, Zhuoyi
    Yuan, Qianmu
    Chen, Sheng
    Yu, Weijiang
    Lu, Yutong
    Gao, Jianzhao
    Yang, Yuedong
    BIOINFORMATICS, 2023, 39 (04)