Semantics of Multiword Expressions in Transformer-Based Models: A Survey

被引:0
|
作者
Miletic, Filip [1 ]
Walde, Sabine Schulte Im [1 ]
机构
[1] Univ Stuttgart, Inst Nat Language Proc, Stuttgart, Germany
关键词
88;
D O I
10.1162/tacl_a_00657
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multiword expressions (MWEs) are composed of multiple words and exhibit variable degrees of compositionality. As such, their meanings are notoriously difficult to model, and it is unclear to what extent this issue affects transformer architectures. Addressing this gap, we provide the first in-depth survey of MWE processing with transformer models. We overall find that they capture MWE semantics inconsistently, as shown by reliance on surface patterns and memorized information. MWE meaning is also strongly localized, predominantly in early layers of the architecture. Representations benefit from specific linguistic properties, such as lower semantic idiosyncrasy and ambiguity of target expressions. Our findings overall question the ability of transformer models to robustly capture fine-grained semantics. Furthermore, we highlight the need for more directly comparable evaluation setups.
引用
收藏
页码:593 / 612
页数:20
相关论文
共 50 条
  • [1] A Distributional account of the semantics of multiword expressions
    Fazly, Afsaneh
    Stevenson, Suzanne
    [J]. ITALIAN JOURNAL OF LINGUISTICS, 2008, 20 (01): : 157 - 179
  • [2] Applications of transformer-based language models in bioinformatics: a survey
    Zhang, Shuang
    Fan, Rui
    Liu, Yuti
    Chen, Shuang
    Liu, Qiao
    Zeng, Wanwen
    [J]. NEURO-ONCOLOGY ADVANCES, 2023, 5 (01)
  • [3] AMMU: A survey of transformer-based biomedical pretrained language models
    Kalyan, Katikapalli Subramanyam
    Rajasekharan, Ajit
    Sangeetha, Sivanesan
    [J]. JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 126
  • [4] Transformer-based language models for mental health issues: A survey
    Greco, Candida M.
    Simeri, Andrea
    Tagarelli, Andrea
    Zumpano, Ester
    [J]. PATTERN RECOGNITION LETTERS, 2023, 167 : 204 - 211
  • [5] Models of Language and Multiword Expressions
    Kallens, Pablo Contreras
    Christiansen, Morten H.
    [J]. FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 5
  • [6] Semanformer: Semantics-aware Embedding Dimensionality Reduction Using Transformer-Based Models
    Boyapati, Mallika
    Aygun, Ramazan
    [J]. 18TH IEEE INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING, ICSC 2024, 2024, : 134 - 141
  • [7] EEG Classification with Transformer-Based Models
    Sun, Jiayao
    Xie, Jin
    Zhou, Huihui
    [J]. 2021 IEEE 3RD GLOBAL CONFERENCE ON LIFE SCIENCES AND TECHNOLOGIES (IEEE LIFETECH 2021), 2021, : 92 - 93
  • [8] Parsing Models for Identifying Multiword Expressions
    Green, Spence
    de Marneffe, Marie-Catherine
    Manning, Christopher D.
    [J]. COMPUTATIONAL LINGUISTICS, 2013, 39 (01) : 195 - 227
  • [9] SentiBERT: A Transferable Transformer-Based Architecture for Compositional Sentiment Semantics
    Yin, Da
    Meng, Tao
    Chang, Kai-Wei
    [J]. 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3695 - 3706
  • [10] Survey of Transformer-Based Object Detection Algorithms
    Li, Jian
    Du, Jianqiang
    Zhu, Yanchen
    Guo, Yongkun
    [J]. Computer Engineering and Applications, 2023, 59 (10) : 48 - 64