On Conditional and Compositional Language Model Differentiable Prompting

被引:0
|
作者
Pilault, Jonathan [1 ]
Liu, Can [2 ]
Bansal, Mohit [3 ]
Dreyer, Markus [2 ]
机构
[1] Polytech Montreal, Mila Quebec AI Inst, Montreal, PQ, Canada
[2] Amazon Alexa, Seattle, WA USA
[3] Univ North Carolina Chapel Hill, Chapel Hill, NC USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) to perform well on downstream tasks. Prompts can be represented by a human-engineered word sequence or by a learned continuous embedding. In this work, we investigate conditional and compositional differentiable prompting. We propose a new model, Prompt Production System (PROPS), which learns to transform task instructions or input metadata, into continuous prompts that elicit task-specific outputs from the PLM. Our model uses a modular network structure based on our neural formulation of Production Systems, which allows the model to learn discrete rules - neural functions that learn to specialize in transforming particular prompt input patterns, making it suitable for compositional transfer learning and few-shot learning. We present extensive empirical and theoretical analysis and show that PROPS consistently surpasses other PLM adaptation techniques, and often improves upon fully fine-tuned models, on compositional generalization tasks, controllable summarization and multilingual translation, while needing fewer trainable parameters.
引用
收藏
页码:4136 / 4144
页数:9
相关论文
共 50 条
  • [41] Graph Neural Prompting with Large Language Models
    Tian, Yijun
    Song, Huan
    Wang, Zichen
    Wang, Haozhu
    Hu, Ziqing
    Wang, Fang
    Chawla, Nitesh V.
    Xu, Panpan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19080 - 19088
  • [42] On the Evolution of Compositional Language
    Barrett, Jeffrey A.
    Cochran, Calvin
    Skyrms, Brian
    PHILOSOPHY OF SCIENCE, 2020, 87 (05) : 910 - 920
  • [43] Prompting Large Language Models With the Socratic Method
    Chang, Edward Y.
    2023 IEEE 13TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE, CCWC, 2023, : 351 - 360
  • [44] Universal Sentence Representation Learning with Conditional Masked Language Model
    Yang, Ziyi
    Yang, Yinfei
    Cer, Daniel
    Law, Jax
    Darve, Eric
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 6216 - 6228
  • [45] A Universal Prompting Strategy for Extracting Process Model Information from Natural Language Text Using Large Language Models
    Neuberger, Julian
    Ackermann, Lars
    van der Aa, Han
    Jablonski, Stefan
    CONCEPTUAL MODELING, ER 2024, 2025, 15238 : 38 - 55
  • [46] Prompting large language model with context and pre-answer for knowledge-based VQA
    Hu, Zhongjian
    Yang, Peng
    Jiang, Yuanshuang
    Bai, Zijian
    PATTERN RECOGNITION, 2024, 151
  • [47] DLAP: A Deep Learning Augmented Large Language Model Prompting framework for software vulnerability detection
    Yang, Yanjing
    Zhou, Xin
    Mao, Runfeng
    Xu, Jinwei
    Yang, Lanxin
    Zhang, Yu
    Shen, Haifeng
    Zhang, He
    JOURNAL OF SYSTEMS AND SOFTWARE, 2025, 219
  • [48] LMCAP: Few-shot Multilingual Image Captioning by Retrieval Augmented Language Model Prompting
    Ramos, Rita
    Martins, Bruno
    Elliott, Desmond
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1635 - 1651
  • [49] Few-shot Reranking for Multi-hop QA via Language Model Prompting
    Khalifa, Muhammad
    Logeswaran, Lajanugen
    Lee, Moontae
    Lee, Honglak
    Wang, Lu
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15882 - 15897
  • [50] Imperative as conditional: From constructional to compositional semantics
    Fortuin, Egbert
    Boogaart, Ronny
    COGNITIVE LINGUISTICS, 2009, 20 (04) : 641 - 673