Eliciting knowledge from language models with automatically generated continuous prompts

被引:1
|
作者
Chen, Yadang [1 ,2 ]
Yang, Gang [1 ,2 ]
Wang, Duolin [1 ,2 ]
Li, Dichao [3 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp Sci, Nanjing 210044, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Engn Res Ctr Digital Forens, Minist Educ, Nanjing 210044, Peoples R China
[3] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
关键词
Prompt learning; Initialization; Trigger token; Continuous parameters;
D O I
10.1016/j.eswa.2023.122327
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained Language Models (PLMs) have demonstrated remarkable performance in Natural Language Under-standing (NLU) tasks, with continuous prompt-based fine-tuning further enhancing their capabilities. However, current methods rely on hand-crafted discrete prompts to initialize continuous prompts, which are sensitive to subtle changes and inherently limited by the constraints of natural language. To address these limitations, this study introduces an innovative AutoPrompt-based Prompt Tuning (APT) approach. APT optimizes the initialization of continuous prompts by employing a gradient-guided automatic search to generate ideal discrete templates and identify trigger tokens. As the semantic features are already captured from the target task dataset, the continuous parameters initialized by trigger tokens are highly relevant, providing a superior starting point for prompt-tuning. APT searches for optimal prompts across various NLU tasks, enabling the PLM to learn task-related knowledge effectively. The APT method significantly improves PLM performance in both few-shot and fully supervised settings, eliminating the need for extensive prompt engineering. In the knowledge exploration (Language Model Analysis (LAMA)) benchmark, APT achieved a remarkable 58.6% (P@1) performance without additional text, representing a 3.6% improvement over the previous best result. Additionally, APT outperformed state-of-the-art methods in the SuperGLUE benchmark.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Advances in the simplification of Fault Trees automatically generated from AltaRica 3.0 models
    Batteux, M.
    Prosvirnova, T.
    Rauzy, A.
    SAFETY AND RELIABILITY - SAFE SOCIETIES IN A CHANGING WORLD, 2018, : 907 - 914
  • [22] Improving Reproducibility of Automatically Generated Markov Models from Molecular Dynamics Trajectories
    Volkhardt, Andreas
    Kozlowski, Nicolai
    Schaeffner, Malte
    Grubmuller, Helmut
    BIOPHYSICAL JOURNAL, 2021, 120 (03) : 79A - 79A
  • [24] ArgumentPrompt: Activating Multi-category of Information for Event Argument Extraction with Automatically Generated Prompts
    Dong, Shenpo
    Yu, Wei
    Tu, Hongkui
    Wang, Xiaodong
    Zhou, Yunyan
    Li, Haili
    Zhou, Jie
    Chang, Tao
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 311 - 323
  • [25] Validation of PLC Code by automatically generated simulation models
    Schmidt, Philipp Puntel
    Fay, Alexander
    Riediger, Willi
    Schulte, Thomas
    Koeslin, Fabian
    Diehl, Stephan
    AT-AUTOMATISIERUNGSTECHNIK, 2015, 63 (02) : 111 - 120
  • [26] SUPPORT OF THE SYSTEM INTEGRATION WITH AUTOMATICALLY GENERATED BEHAVIOUR MODELS
    Koessler, Johannes
    Paetzold, Kristin
    ICED 15, VOL 11: HUMAN BEHAVIOUR IN DESIGN, DESIGN EDUCATION, 2015,
  • [27] Failure prediction in automatically generated digital elevation models
    Gooch, MJ
    Chandler, JH
    COMPUTERS & GEOSCIENCES, 2001, 27 (08) : 913 - 920
  • [28] Eliciting Offensive Responses from Large Language Models: A Genetic Algorithm Approach
    Chen, Zheng
    Zhu, Jiachen
    Chen, Anlong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14864 : 456 - 467
  • [29] How to write effective prompts for large language models
    Lin, Zhicheng
    NATURE HUMAN BEHAVIOUR, 2024, 8 (4) : 611 - 615
  • [30] How to write effective prompts for large language models
    Zhicheng Lin
    Nature Human Behaviour, 2024, 8 : 611 - 615