Eliciting knowledge from language models with automatically generated continuous prompts

被引:1
|
作者
Chen, Yadang [1 ,2 ]
Yang, Gang [1 ,2 ]
Wang, Duolin [1 ,2 ]
Li, Dichao [3 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp Sci, Nanjing 210044, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Engn Res Ctr Digital Forens, Minist Educ, Nanjing 210044, Peoples R China
[3] Nanjing Univ Informat Sci & Technol, Sch Artificial Intelligence, Nanjing 210044, Peoples R China
关键词
Prompt learning; Initialization; Trigger token; Continuous parameters;
D O I
10.1016/j.eswa.2023.122327
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained Language Models (PLMs) have demonstrated remarkable performance in Natural Language Under-standing (NLU) tasks, with continuous prompt-based fine-tuning further enhancing their capabilities. However, current methods rely on hand-crafted discrete prompts to initialize continuous prompts, which are sensitive to subtle changes and inherently limited by the constraints of natural language. To address these limitations, this study introduces an innovative AutoPrompt-based Prompt Tuning (APT) approach. APT optimizes the initialization of continuous prompts by employing a gradient-guided automatic search to generate ideal discrete templates and identify trigger tokens. As the semantic features are already captured from the target task dataset, the continuous parameters initialized by trigger tokens are highly relevant, providing a superior starting point for prompt-tuning. APT searches for optimal prompts across various NLU tasks, enabling the PLM to learn task-related knowledge effectively. The APT method significantly improves PLM performance in both few-shot and fully supervised settings, eliminating the need for extensive prompt engineering. In the knowledge exploration (Language Model Analysis (LAMA)) benchmark, APT achieved a remarkable 58.6% (P@1) performance without additional text, representing a 3.6% improvement over the previous best result. Additionally, APT outperformed state-of-the-art methods in the SuperGLUE benchmark.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Can Large Language Models Truly Understand Prompts? A Case Study with Negated Prompts
    Jang, Joel
    Ye, Seongheyon
    Seo, Minjoon
    TRANSFER LEARNING FOR NATURAL LANGUAGE PROCESSING WORKSHOP, VOL 203, 2022, 203 : 52 - 62
  • [32] Efficient Detection of Toxic Prompts in Large Language Models
    Liu, Yi
    Yu, Junzhe
    Sun, Huijia
    Shi, Ling
    Deng, Gelei
    Chen, Yuqi
    Liu, Yang
    arXiv, 1600,
  • [33] Demystifying Prompts in Language Models via Perplexity Estimation
    Gonen, Hila
    Iyer, Srini
    Blevins, Terra
    Smith, Noah A.
    Zettlemoyer, Luke
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 10136 - 10148
  • [34] A MULTI-LANGUAGE COMPILER SYSTEM WITH AUTOMATICALLY GENERATED CODEGENERATORS
    SCHMIDT, U
    VOLLER, R
    SIGPLAN NOTICES, 1984, 19 (06): : 202 - 212
  • [35] Therapy Language Analysis using Automatically Generated Psycholinguistic Norms
    Malandrakis, Nikolaos
    Narayanan, Shrikanth
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1952 - 1956
  • [36] INCREASING MODULARITY AND LANGUAGE-INDEPENDENCY IN AUTOMATICALLY GENERATED COMPILERS
    GANZINGER, H
    SCIENCE OF COMPUTER PROGRAMMING, 1983, 3 (03) : 223 - 278
  • [37] Was the title of this talk generated automatically? Prospects on intelligent interfaces and language
    Stock, O
    IJCAI-99: PROCEEDINGS OF THE SIXTEENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOLS 1 & 2, 1999, : 1412 - 1419
  • [38] Generality Evaluation of Automatically Generated Knowledge for the Japanese Concept Net
    Rzepka, Rafal
    Muramoto, Koichi
    Araki, Kenji
    AI 2011: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2011, 7106 : 648 - 657
  • [39] AK4Prompts: Aesthetics-driven Automatically Keywords-Ranking for Prompts in Text-To-Image Models
    Zhang, Haiyang
    Wang, Mengchao
    He, Shuai
    Ming, Anlong
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 1661 - 1669
  • [40] CLIPSwarm: Generating Drone Shows from Text Prompts with Vision-Language Models
    Pueyo, Pablo
    Montijano, Eduardo
    Murillo, Ana C.
    Mac Schwager
    2024 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2024), 2024, : 11917 - 11923