A Multi-Task Semantic Decomposition Framework with Task-specific Pre-training for Few-Shot NER

被引:7
|
作者
Dong, Guanting [1 ]
Wang, Zechen [1 ]
Zhao, Jinxu [1 ]
Zhao, Gang [1 ]
Guo, Daichi [1 ]
Fu, Dayuan [1 ]
Hui, Tingfeng [1 ]
Zeng, Chen [1 ]
He, Keqing [2 ]
Li, Xuefeng [1 ]
Wang, Liwen [1 ]
Cui, Xinyue [1 ]
Xu, Weiran [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
[2] Meituan Grp Beijing, Beijing, Peoples R China
关键词
Few-shot NER; Multi-Task; Semantic Decomposition; Pre-training;
D O I
10.1145/3583780.3614766
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The objective of few-shot named entity recognition is to identify named entities with limited labeled instances. Previous works have primarily focused on optimizing the traditional token-wise classification framework, while neglecting the exploration of information based on NER data characteristics. To address this issue, we propose a Multi-Task Semantic Decomposition Framework via Joint Task-specific Pre-training (MSDP) for few-shot NER. Drawing inspiration from demonstration-based and contrastive learning, we introduce two novel pre-training tasks: Demonstration-based Masked Language Modeling (MLM) and Class Contrastive Discrimination. These tasks effectively incorporate entity boundary information and enhance entity representation in Pre-trained Language Models (PLMs). In the downstream main task, we introduce a multitask joint optimization framework with the semantic decomposing method, which facilitates the model to integrate two different semantic information for entity classification. Experimental results of two few-shot NER benchmarks demonstrate that MSDP consistently outperforms strong baselines by a large margin. Extensive analyses validate the effectiveness and generalization of MSDP.
引用
收藏
页码:430 / 440
页数:11
相关论文
共 50 条
  • [31] A Multi-Task Instruction with Chain of Thought Prompting Generative Framework for Few-Shot Named Entity Recognition
    Xu, WenJie
    OuYang, JianQuan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT V, 2023, 14258 : 1 - 15
  • [32] Improving Few-Shot Learning Through Multi-task Representation Learning Theory
    Bouniot, Quentin
    Redko, Ievgen
    Audigier, Romaric
    Loesch, Angelique
    Habrard, Amaury
    COMPUTER VISION, ECCV 2022, PT XX, 2022, 13680 : 435 - 452
  • [33] Multi-task few-shot learning with composed data augmentation for image classification
    Zhang, Rui
    Yang, Yixin
    Li, Yang
    Wang, Jiabao
    Li, Hang
    Miao, Zhuang
    IET COMPUTER VISION, 2023, 17 (02) : 211 - 221
  • [34] SIM: an improved few-shot image classification model with multi-task learning
    Guo, Jin
    Li, Wengen
    Guan, Jihong
    Gao, Hang
    Liu, Baobo
    Gong, Lili
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (03)
  • [35] Task-specific contrastive learning for few-shot remote sensing image scene classification
    Zeng, Qingjie
    Geng, Jie
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2022, 191 : 143 - 154
  • [36] Learning Task-Specific Embeddings for Few-Shot Classification via Local Weight Adaptation
    Gong, Nianru
    Duan, Pengfei
    Rong, Yi
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 485 - 491
  • [37] Task-Specific Alignment and Multiple-level Transformer for few-shot action recognition
    Guo, Fei
    Zhu, Li
    Wang, Yikang
    Sun, Jing
    NEUROCOMPUTING, 2024, 598
  • [38] MLDAC:Multi-Task Dense Attention Computation Self-Supervised Few-Shot Semantic Segmentation Method
    Wang, Weihang
    Zhang, Yi
    Computer Engineering and Applications, 2025, 61 (04) : 211 - 221
  • [39] SNIP-FSL: Finding task-specific lottery jackpots for few-shot learning
    Wang, Ren
    Sun, Haoliang
    Nie, Xiushan
    Yin, Yilong
    KNOWLEDGE-BASED SYSTEMS, 2022, 247
  • [40] A Kernel Approach to Multi-Task Learning with Task-Specific Kernels
    Wu, Wei
    Li, Hang
    Hu, Yun-Hua
    Jin, Rong
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2012, 27 (06) : 1289 - 1301