Reinforced Iterative Knowledge Distillation for Cross-Lingual Named Entity Recognition

被引:13
|
作者
Liang, Shining [1 ,2 ,5 ]
Gong, Ming [3 ]
Pei, Jian [4 ]
Shou, Linjun [3 ]
Zuo, Wanli [1 ,2 ]
Zuo, Xianglin [1 ,2 ]
Jiang, Daxin [3 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Beijing, Peoples R China
[3] Microsoft, STCA NLP Grp, Beijing, Peoples R China
[4] Simon Fraser Univ, Sch Comp Sci, Burnaby, BC, Canada
[5] Microsoft STCA, Beijing, Peoples R China
基金
加拿大自然科学与工程研究理事会; 中国国家自然科学基金;
关键词
named entity recognition; knowledge distillation; reinforcement learning; cross lingual;
D O I
10.1145/3447548.3467196
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Named entity recognition (NER) is a fundamental component in many applications, such as Web Search and Voice Assistants. Although deep neural networks greatly improve the performance of NER, due to the requirement of large amounts of training data, deep neural networks can hardly scale out to many languages in an industry setting. To tackle this challenge, cross-lingual NER transfers knowledge from a rich-resource language to languages with low resources through pre-trained multilingual language models. Instead of using training data in target languages, cross-lingual NER has to rely on only training data in source languages, and optionally adds the translated training data derived from source languages. However, the existing cross-lingual NER methods do not make good use of rich unlabeled data in target languages, which is relatively easy to collect in industry applications. To address the opportunities and challenges, in this paper we describe our novel practice in Microsoft to leverage such large amounts of unlabeled data in target languages in real production settings. To effectively extract weak supervision signals from the unlabeled data, we develop a novel approach based on the ideas of semi-supervised learning and reinforcement learning. The empirical study on three benchmark data sets verifies that our approach establishes the new state-of-the-art performance with clear edges. Now, the NER techniques reported in this paper are on their way to become a fundamental component for Web ranking, Entity Pane, Answers Triggering, and Question Answering in the Microsoft Bing search engine. Moreover, our techniques will also serve as part of the Spoken Language Understanding module for a commercial voice assistant. We plan to open source the code of the prototype framework after deployment.
引用
收藏
页码:3231 / 3239
页数:9
相关论文
共 50 条
  • [1] Cross-lingual Named Entity Recognition
    Steinberger, Ralf
    Pouliquen, Bruno
    [J]. LINGUISTICAE INVESTIGATIONES, 2007, 30 (01): : 135 - 162
  • [2] WASSERSTEIN CROSS-LINGUAL ALIGNMENT FOR NAMED ENTITY RECOGNITION
    Wang, Rui
    Henao, Ricardo
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8342 - 8346
  • [3] Cross-Lingual Named Entity Recognition for Heterogenous Languages
    Fu, Yingwen
    Lin, Nankai
    Chen, Boyu
    Yang, Ziyu
    Jiang, Shengyi
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 371 - 382
  • [4] Discrepancy and Uncertainty Aware Denoising Knowledge Distillation for Zero-Shot Cross-Lingual Named Entity Recognition
    Ge, Ling
    Hu, Chunming
    Ma, Guanghui
    Liu, Jihong
    Zhang, Hong
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18056 - 18064
  • [5] Exploiting Morpheme and Cross-lingual Knowledge to Enhance Mongolian Named Entity Recognition
    Zhang, Songming
    Zhang, Ying
    Chen, Yufeng
    Wu, Du
    Xu, Jinan
    Liu, Jian
    [J]. ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2022, 21 (05)
  • [6] Improving Low Resource Named Entity Recognition using Cross-lingual Knowledge Transfer
    Feng, Xiaocheng
    Feng, Xiachong
    Qin, Bing
    Feng, Zhangyin
    Liu, Ting
    [J]. PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 4071 - 4077
  • [7] Neural Cross-Lingual Named Entity Recognition with Minimal Resources
    Xie, Jiateng
    Yang, Zhilin
    Neubig, Graham
    Smith, Noah A.
    Carbonell, Jaime
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 369 - 379
  • [8] Cross-lingual Transfer Learning for Japanese Named Entity Recognition
    Johnson, Andrew
    Karanasou, Penny
    Gaspers, Judith
    Klakow, Dietrich
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES(NAACL HLT 2019), VOL. 2 (INDUSTRY PAPERS), 2019, : 182 - 189
  • [9] Zero-Resource Cross-Lingual Named Entity Recognition
    Bari, M. Saiful
    Joty, Shafiq
    Jwalapuram, Prathyusha
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7415 - 7423
  • [10] Cross-Lingual Transfer Learning for Medical Named Entity Recognition
    Ding, Pengjie
    Wang, Lei
    Liang, Yaobo
    Lu, Wei
    Li, Linfeng
    Wang, Chun
    Tang, Buzhou
    Yan, Jun
    [J]. DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2020), PT I, 2020, 12112 : 403 - 418