Cross-Lingual Named Entity Recognition for Heterogenous Languages

被引:1
|
作者
Fu, Yingwen [1 ]
Lin, Nankai [2 ]
Chen, Boyu [3 ]
Yang, Ziyu [1 ]
Jiang, Shengyi [1 ,4 ]
机构
[1] Guangdong Univ Foreign Studies, Sch Informat Sci & Technol, Guangzhou 510006, Guangdong, Peoples R China
[2] Guangdong Univ Technol, Sch Comp Sci & Technol, Guangzhou 510006, Guangdong, Peoples R China
[3] UCL, Inst Hlth Informat, London WC1E 6BT, England
[4] Guangdong Univ Foreign Studies, Guangzhou Key Lab Multilingual Intelligent Proc, Guangzhou 510006, Guangdong, Peoples R China
关键词
Training; Data models; Standards; Speech processing; Optimization; Knowledge transfer; Information science; Cross-lingual named entity recognition; heterogenous language; weakly supervised learning; bilateral-branch network; self-distillation;
D O I
10.1109/TASLP.2022.3212698
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Previous works on cross-lingual Named Entity Recognition (NER) have achieved great success. However, few of them consider the effect of language families between the source and target languages. In this study, we find that the cross-lingual NER performance of a target language would decrease when its source language is changed from the same (homogenous) into a different (heterogenous) language family with that target language. To improve the NER performance in this situation, we propose a novel cross-lingual NER framework based on self-distillation mechanism and Bilateral-Branch Network (SD-BBN). SD-BBN learns source-language NER knowledge from supervised datasets and obtains target-language knowledge from weakly supervised datasets. These two kinds of knowledge are then fused based on self-distillation mechanism for better identifying entities in the target language. We evaluate SD-BBN on 9 language datasets from 4 different language families. Results show that SD-BBN tends to outperform baseline methods. Remarkably, when the target and source languages are heterogenous, SD-BBN can achieve a greater boost. Our results might suggest that obtaining language-specific knowledge from the target language is essential for improving cross-lingual NER when the source and target languages are heterogenous. This finding could provide a novel insight into further research.
引用
下载
收藏
页码:371 / 382
页数:12
相关论文
共 50 条
  • [31] Cross-Lingual Cross-Domain Nested Named Entity Evaluation on EnglishWeb Texts
    Plank, Barbara
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1808 - 1815
  • [32] Discrepancy and Uncertainty Aware Denoising Knowledge Distillation for Zero-Shot Cross-Lingual Named Entity Recognition
    Ge, Ling
    Hu, Chunming
    Ma, Guanghui
    Liu, Jihong
    Zhang, Hong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18056 - 18064
  • [33] Zero-Shot Cross-Lingual Named Entity Recognition via Progressive Multi-Teacher Distillation
    Li, Zhuoran
    Hu, Chunming
    Zhang, Richong
    Chen, Junfan
    Guo, Xiaohui
    IEEE/ACM Transactions on Audio Speech and Language Processing, 2024, 32 : 4617 - 4630
  • [34] Low Resource Named Entity Recognition Using Contextual Word Representation and Neural Cross-Lingual Knowledge Transfer
    Han, Soyeon Caren
    Lin, Yingru
    Long, Siqu
    Poon, Josiah
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 299 - 311
  • [35] Neural Cross-Lingual Entity Linking
    Sil, Avirup
    Kundu, Gourab
    Florian, Radu
    Hamza, Wael
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 5464 - 5472
  • [36] Cross-Lingual UMLS Named Entity Linking using UMLS Dictionary Fine-Tuning
    Galperin, Rina
    Schnapp, Shachar
    Elhadad, Michael
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3380 - 3390
  • [37] Cross-Lingual Entity Linking for Web Tables
    Luo, Xusheng
    Luo, Kangqi
    Chen, Xianyang
    Zhu, Kenny Q.
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 362 - 369
  • [38] Cross-lingual Entity Alignment with Incidental Supervision
    Chen, Muhao
    Shi, Weijia
    Zhou, Ben
    Roth, Dan
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 645 - 658
  • [39] Cross-Lingual Entity Linking in Wikipedia Infoboxes
    Yang, Juheng
    Wang, Zhichun
    KNOWLEDGE GRAPH AND SEMANTIC COMPUTING: KNOWLEDGE COMPUTING AND LANGUAGE UNDERSTANDING, 2019, 1134 : 38 - 49
  • [40] CROSS-LINGUAL SPEECH RECOGNITION BETWEEN LANGUAGES FROM THE SAME LANGUAGE FAMILY
    Zgank, Andrej
    PROCEEDINGS OF THE ROMANIAN ACADEMY SERIES A-MATHEMATICS PHYSICS TECHNICAL SCIENCES INFORMATION SCIENCE, 2019, 20 (02): : 184 - 191