MsPrompt: Multi-step prompt learning for debiasing few-shot event detection

被引:6
|
作者
Wang, Siyuan [1 ]
Zheng, Jianming [1 ]
Cai, Fei [1 ]
Song, Chengyu [1 ]
Luo, Xueshan [1 ]
机构
[1] Natl Univ Def Technol, Sci & Technol Informat Syst Engn Lab, Changsha, Hunan, Peoples R China
基金
中国国家自然科学基金;
关键词
Few-shot event detection; Prompt learning; Debiasing;
D O I
10.1016/j.ipm.2023.103509
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Event detection (ED) is aimed to identify the key trigger words in unstructured text and predict the event types accordingly. Traditional ED models are too data-hungry to accommodate real applications with scarce labeled data. Besides, typical ED models are facing the context -bypassing and disabled generalization issues caused by the trigger bias stemming from ED datasets. Therefore, we focus on the true few-shot paradigm targeting to construct a novel training set that accommodates the low-resource scenarios. In particular, we propose a multi-step prompt learning model (MsPrompt) for debiasing few-shot event detection (FSED), consisting of the following two components: a multi-step prompt module equipped with a knowledge-enhanced ontology to leverage the event semantics and latent prior knowledge in the pretrained language models (PLMs) sufficiently for tackling the context-bypassing problem, and a prototypical network module compensating for the weakness of classifying events with sparse data and boost the generalization performance. Experiments on two public datasets ACE-2005 and FewEvent show that MsPrompt can outperform the state-of-the-art models, especially in the strict low-resource scenarios reporting 11.43% improvement in terms of weighted F1-score against the best baseline and achieving an outstanding debiasing performance.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Few-Shot Learning with Novelty Detection
    Bjerge, Kim
    Bodesheim, Paul
    Karstoft, Henrik
    DEEP LEARNING THEORY AND APPLICATIONS, PT I, DELTA 2024, 2024, 2171 : 340 - 363
  • [22] ID-like Prompt Learning for Few-Shot Out-of-Distribution Detection
    Bai, Yichen
    Han, Zongbo
    Cao, Bing
    Jiang, Xiaoheng
    Hu, Qinghua
    Zhang, Changqing
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 17480 - 17489
  • [23] VulPrompt: Prompt-Based Vulnerability Detection Using Few-Shot Graph Learning
    Irtiza, Saquib
    Li, Xiaodi
    Zamani, Mahmoud
    Khan, Latifur
    Hamlen, Kevin W.
    DATA AND APPLICATIONS SECURITY AND PRIVACY XXXVIII, DBSEC 2024, 2024, 14901 : 221 - 240
  • [24] Few-Shot Composition Learning for Image Retrieval with Prompt Tuning
    Wu, Junda
    Wang, Rui
    Zhao, Handong
    Zhang, Ruiyi
    Lu, Chaochao
    Li, Shuai
    Henao, Ricardo
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 4, 2023, : 4729 - 4737
  • [25] Enhanced Prompt Learning for Few-shot Text Classification Method
    Li R.
    Wei Z.
    Fan Y.
    Ye S.
    Zhang G.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2024, 60 (01): : 1 - 12
  • [26] Prompt-Based Metric Learning for Few-Shot NER
    Chen, Yanru
    Zheng, Yanan
    Yang, Zhilin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 7199 - 7212
  • [27] Log Parsing with Prompt-based Few-shot Learning
    Le, Van-Hoang
    Zhang, Hongyu
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ICSE, 2023, : 2438 - 2449
  • [28] Learning Meta Soft Prompt for Few-Shot Language Models
    Chien, Jen-Tzung
    Chen, Ming-Yen
    Xue, Jing-Hao
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 57 - 62
  • [29] Multi-channels Prototype Contrastive Learning with Condition Adversarial Attacks for Few-shot Event Detection
    Zhang, Fangchen
    Tian, Shengwei
    Yu, Long
    Yang, Qimeng
    NEURAL PROCESSING LETTERS, 2024, 56 (01)
  • [30] Multi-channels Prototype Contrastive Learning with Condition Adversarial Attacks for Few-shot Event Detection
    Fangchen Zhang
    Shengwei Tian
    Long Yu
    Qimeng Yang
    Neural Processing Letters, 56