CDANER: Contrastive Learning with Cross-domain Attention for Few-shot Named Entity Recognition

被引:2
|
作者
Li, Wei [1 ,2 ]
Li, Hui [2 ]
Ge, Jingguo [1 ,2 ]
Zhang, Lei [2 ]
Li, Liangxiong [2 ]
Wu, Bingzhen [2 ]
机构
[1] Univ Chinese Acad Sci, Sch Cyberspace Secur, Beijing, Peoples R China
[2] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
基金
国家重点研发计划;
关键词
D O I
10.1109/IJCNN54540.2023.10191439
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot Named Entity Recognition (NER) aims to recognize unseen name entities based on a tiny support set that consists of seen name entities and labels, which is obviously different from traditional supervised NER methods. Contrastive learning has become a popular solution for few-shot NER, which improves the robustness of NER to handle unlabeled entities by learning a similarity metric to measure the semantic similarity between test samples and entity labels. However, existing contrastive learning based NER methods individually learn the word embedding in source and target domains, ignoring connections between entities with the same label and limiting the effectiveness of contrast learning. In this paper, we propose a novel few-shot NER framework that jointly models different domain texts and optimizes a generalized objective of differentiating between words in all stages. The proposed model builds the cross-domain attention layer to enhance the feature representations of words and transfer the entity similarity information from the source domain to the target domain. This significantly reduces the divergence between entities with same label. Experimental results on the largest Few-shot NER dataset show that CDANER significantly outperforms all baseline methods, which verifies the effectiveness and robustness of the proposed model.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] CONTAINER: Few-Shot Named Entity Recognition via Contrastive Learning
    Das, Sarkar Snigdha Sarathi
    Katiyar, Arzoo
    Passonneau, Rebecca J.
    Zhang, Rui
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 6338 - 6353
  • [2] Dual Contrastive Learning for Cross-Domain Named Entity Recognition
    Xu, Jingyun
    Yu, Junnan
    Cai, Yi
    Chua, Tat-Seng
    [J]. ACM Transactions on Information Systems, 2024, 42 (06)
  • [3] Cross-Domain Few-Shot Contrastive Learning for Hyperspectral Images Classification
    Zhang, Suhua
    Chen, Zhikui
    Wang, Dan
    Wang, Z. Jane
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [4] Cross-Domain Contrastive Learning-Based Few-Shot Underwater Acoustic Target Recognition
    Cui, Xiaodong
    He, Zhuofan
    Xue, Yangtao
    Tang, Keke
    Zhu, Peican
    Han, Jing
    [J]. JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2024, 12 (02)
  • [5] Meta-Learning for Few-Shot Named Entity Recognition
    de Lichy, Cyprien
    Glaude, Hadrien
    Campbell, William
    [J]. 1ST WORKSHOP ON META LEARNING AND ITS APPLICATIONS TO NATURAL LANGUAGE PROCESSING (METANLP 2021), 2021, : 44 - 58
  • [6] Prompts in Few-Shot Named Entity Recognition
    Rozhkov, I. S.
    Loukachevitch, N. V.
    [J]. PATTERN RECOGNITION AND IMAGE ANALYSIS, 2023, 33 (02) : 122 - 131
  • [7] Few-shot nested named entity recognition
    Ming, Hong
    Yang, Jiaoyun
    Gui, Fang
    Jiang, Lili
    An, Ning
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 293
  • [8] Prompts in Few-Shot Named Entity Recognition
    I. S. Rozhkov
    N. V. Loukachevitch
    [J]. Pattern Recognition and Image Analysis, 2023, 33 : 122 - 131
  • [9] Understanding Cross-Domain Few-Shot Learning Based on Domain Similarity and Few-Shot Difficulty
    Oh, Jaehoon
    Kim, Sungnyun
    Ho, Namgyu
    Kim, Jin-Hwa
    Song, Hwanjun
    Yun, Se-Young
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [10] Attending to Entity Class Attributes for Named Entity Recognition with Few-Shot Learning
    Patel, Raj Nath
    Dutta, Sourav
    Assem, Haytham
    [J]. INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 3, INTELLISYS 2023, 2024, 824 : 859 - 870