Morphosyntactic probing of multilingual BERT models

被引:2
|
作者
Acs, Judit [1 ,2 ]
Hamerlik, Endre [1 ,4 ]
Schwartz, Roy [5 ]
Smith, Noah A. [6 ,7 ]
Kornai, Andras [1 ,3 ]
机构
[1] ELKH Inst Comp Sci & Control SZTAK, Informat Lab, Budapest, Hungary
[2] Budapest Univ Technol & Econ, Fac Elect Engn & Informat, Dept Automat & Appl Informat, Budapest, Hungary
[3] Budapest Univ Technol & Econ, Fac Nat Sci, Dept Algebra, Budapest, Hungary
[4] Comenius Univ, Fac Math Phys & Informat, Dept Appl Informat, Bratislava, Slovakia
[5] Hebrew Univ Jerusalem, Sch Comp Sci & Engn, Jerusalem, Israel
[6] Univ Washington, Paul G Allen Sch Comp Sci & Engn, Seattle, WA USA
[7] Allen Inst Artificial Intelligence, Seattle, WA USA
关键词
Morphology; Language Resources; Multilinguality; Machine Learning; Language Models;
D O I
10.1017/S1351324923000190
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce an extensive dataset for multilingual probing of morphological information in language models (247 tasks across 42 languages from 10 families), each consisting of a sentence with a target word and a morphological tag as the desired label, derived from the Universal Dependencies treebanks. We find that pre-trained Transformer models (mBERT and XLM-RoBERTa) learn features that attain strong performance across these tasks. We then apply two methods to locate, for each probing task, where the disambiguating information resides in the input. The first is a new perturbation method that "masks" various parts of context; the second is the classical method of Shapley values. The most intriguing finding that emerges is a strong tendency for the preceding context to hold more information relevant to the prediction than the following context.
引用
收藏
页码:753 / 792
页数:40
相关论文
共 50 条
  • [41] Syntax-augmented Multilingual BERT for Cross-lingual Transfer
    Ahmad, Wasi Uddin
    Li, Haoran
    Chang, Kai-Wei
    Mehdad, Yashar
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4538 - 4554
  • [42] bert2BERT: Towards Reusable Pretrained Language Models
    Chen, Cheng
    Yin, Yichun
    Shang, Lifeng
    Jiang, Xin
    Qin, Yujia
    Wang, Fengyu
    Wang, Zhi
    Chen, Xiao
    Liu, Zhiyuan
    Liu, Qun
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2134 - 2148
  • [43] When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer
    Deshpande, Ameet
    Talukdar, Partha
    Narasimhan, Karthik
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3610 - 3623
  • [44] Probing Multilingual Sentence Representations With X-PROBE
    Ravishankar, Vinit
    Ovrelid, Lilja
    Velldal, Erik
    4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), 2019, : 156 - 168
  • [45] Multi-class sentiment analysis of urdu text using multilingual BERT
    Khan, Lal
    Amjad, Ammar
    Ashraf, Noman
    Chang, Hsien-Tsung
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [46] Adversarial Domain Adaptation for Cross-lingual Information Retrieval with Multilingual BERT
    Wang, Runchuan
    Zhang, Zhao
    Zhuang, Fuzhen
    Gao, Dehong
    Wei, Yi
    He, Qing
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3498 - 3502
  • [47] Learning to Match Job Candidates Using Multilingual Bi-Encoder BERT
    Lavi, Dor
    15TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS 2021), 2021, : 565 - 566
  • [48] Multi-class sentiment analysis of urdu text using multilingual BERT
    Lal Khan
    Ammar Amjad
    Noman Ashraf
    Hsien-Tsung Chang
    Scientific Reports, 12
  • [49] LINSPECTOR WEB: A Multilingual Probing Suite for Word Representations
    Eichler, Max
    Sahin, Gozde Gul
    Gurevych, Iryna
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF SYSTEM DEMONSTRATIONS, 2019, : 127 - 132
  • [50] First Align, then Predict: Understanding the Cross-Lingual Ability of Multilingual BERT
    Muller, Benjamin
    Elazar, Yanai
    Sagot, Benoit
    Seddah, Djame
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 2214 - 2231