Multi-Mask Label Mapping for Prompt-Based Learning

被引:0
|
作者
Qi, Jirui [1 ]
Zhang, Richong [1 ,2 ]
Kim, Jaein [1 ]
Chen, Junfan [1 ]
Qin, Wenyi [1 ]
Mao, Yongyi [3 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, SKLSDE, Beijing, Peoples R China
[2] Zhongguancun Lab, Beijing, Peoples R China
[3] Univ Ottawa, Sch Elect Engn & Comp Sci, Ottawa, ON, Canada
基金
国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt-based Learning has shown significant success in few-shot classification. The mainstream approach is to concatenate a template for the input text to transform the classification task into a cloze-type task where label mapping plays an important role in finding the ground-truth labels. While cur-rent label mapping methods only use the contexts in one single input, it could be crucial if wrong information is contained in the text. Specifically, it is proved in recent work that even the large language models like BERT/RoBERTa make classification decisions heavily dependent on a specific keyword regardless of the task or the context. Such a word is referred to as a lexical cue and if a misleading lexical cue is included in the instance it will lead the model to make a wrong prediction. We propose a multi-mask prompt-based approach with Multi-Mask Label Mapping (MMLM) to reduce the impact of misleading lexical cues by allowing the model to exploit multiple lexical cues. To satisfy the conditions of few-shot learning, an instance augmentation approach for the cloze-type model is proposed and the misleading cues are gradually excluded through training. We demonstrate the effectiveness of MMLM by both theoretical analysis and empirical studies, and show that MMLM outperforms other existing label mapping approaches.
引用
收藏
页码:13465 / 13473
页数:9
相关论文
共 50 条
  • [41] Joint contrastive learning for prompt-based few-shot language learners
    Zhengzhong Zhu
    Xuejie Zhang
    Jin Wang
    Xiaobing Zhou
    Neural Computing and Applications, 2024, 36 : 7861 - 7875
  • [42] Prompt-based Few-shot Learning for Table-based Fact Verification
    Hou, Lei
    Liu, Yubo
    Wu, Jie
    Hou, Mengshu
    2022 5TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND NATURAL LANGUAGE PROCESSING, MLNLP 2022, 2022, : 14 - 19
  • [43] Joint contrastive learning for prompt-based few-shot language learners
    Zhu, Zhengzhong
    Zhang, Xuejie
    Wang, Jin
    Zhou, Xiaobing
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (14): : 7861 - 7875
  • [44] Co-training Improves Prompt-based Learning for Large Language Models
    Lang, Hunter
    Agrawal, Monica
    Kim, Yoon
    Sontag, David
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [45] COVER: A Heuristic Greedy Adversarial Attack on Prompt-Based Learning in Language Models
    Chen, Qingliang (tpchen@jnu.edu.cn), 1600, Springer Science and Business Media Deutschland GmbH (14326 LNAI):
  • [46] Continual Few-Shot Relation Extraction with Prompt-Based Contrastive Learning
    Wu, Fei
    Zhang, Chong
    Tan, Zhen
    Xu, Hao
    Ge, Bin
    WEB AND BIG DATA, PT IV, APWEB-WAIM 2023, 2024, 14334 : 312 - 327
  • [47] Multi-Mask Fusion-Based RGB-D SLAM in Dynamic Environments
    Gao Y.
    Hu M.
    Chen B.
    Yang W.
    Wang J.
    Wang J.
    IEEE Sensors Journal, 2024, 24 (21) : 1 - 1
  • [48] A Visual Prompt-Based Mobile Learning System for Improved Algebraic Understanding in Students With Learning Disabilities
    Chang, Peng-Chan
    Lin, Rong-Ho
    IEEE ACCESS, 2024, 12 : 3540 - 3553
  • [49] Prompt-Based Editing for Text Style Transfer
    Luo, Guoqing
    Han, Yu Tong
    Mou, Lili
    Firdaus, Mauajama
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 5740 - 5750
  • [50] Knowledge-based dynamic prompt learning for multi-label disease diagnosis
    Xie, Jing
    Li, Xin
    Yuan, Ye
    Guan, Yi
    Jiang, Jingchi
    Guo, Xitong
    Peng, Xin
    KNOWLEDGE-BASED SYSTEMS, 2024, 286