A Communication Theory Perspective on Prompting Engineering Methods for Large Language Models

被引:0
|
作者
Song, Yuan-Feng [1 ]
He, Yuan-Qin [1 ]
Zhao, Xue-Fang [1 ]
Gu, Han-Lin [1 ]
Jiang, Di [1 ]
Yang, Hai-Jun [1 ]
Fan, Li-Xin [1 ]
机构
[1] AI Group, WeBank Co., Ltd, Shenzhen,518000, China
关键词
Contrastive Learning - Modeling languages - Natural language processing systems - Self-supervised learning - Semi-supervised learning - Zero-shot learning;
D O I
10.1007/s11390-024-4058-8
中图分类号
学科分类号
摘要
The springing up of large language models (LLMs) has shifted the community from single-task-orientated natural language processing (NLP) research to a holistic end-to-end multi-task learning paradigm. Along this line of research endeavors in the area, LLM-based prompting methods have attracted much attention, partially due to the technological advantages brought by prompt engineering (PE) as well as the underlying NLP principles disclosed by various prompting methods. Traditional supervised learning usually requires training a model based on labeled data and then making predictions. In contrast, PE methods directly use the powerful capabilities of existing LLMs (e.g., GPT-3 and GPT-4) via composing appropriate prompts, especially under few-shot or zero-shot scenarios. Facing the abundance of studies related to the prompting and the ever-evolving nature of this field, this article aims to 1) illustrate a novel perspective to review existing PE methods within the well-established communication theory framework, 2) facilitate a better/deeper understanding of developing trends of existing PE methods used in three typical tasks, and 3) shed light on promising research directions for future PE methods. © Institute of Computing Technology, Chinese Academy of Sciences 2024.
引用
收藏
页码:984 / 1004
相关论文
共 50 条
  • [1] Considerations for Prompting Large Language Models
    Schulte, Brian
    [J]. JAMA ONCOLOGY, 2024, 10 (04) : 475 - 483
  • [2] Prompting Is Programming: A Query Language for Large Language Models
    Beurer-Kellner, Luca
    Fischer, Marc
    Vechev, Martin
    [J]. PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2023, 7 (PLDI):
  • [3] Dehallucinating Large Language Models Using Formal Methods Guided Iterative Prompting
    Jha, Susmit
    Jha, Sumit Kumar
    Lincoln, Patrick
    Bastian, Nathaniel D.
    Velasquez, Alvaro
    Neema, Sandeep
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ASSURED AUTONOMY, ICAA, 2023, : 149 - 152
  • [4] Graph Neural Prompting with Large Language Models
    Tian, Yijun
    Song, Huan
    Wang, Zichen
    Wang, Haozhu
    Hu, Ziqing
    Wang, Fang
    Chawla, Nitesh V.
    Xu, Panpan
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19080 - 19088
  • [5] Prompting Large Language Models With the Socratic Method
    Chang, Edward Y.
    [J]. 2023 IEEE 13TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE, CCWC, 2023, : 351 - 360
  • [6] Prompting Large Language Models to Power Educational Chatbots
    Farah, Juan Carlos
    Ingram, Sandy
    Spaenlehauer, Basile
    Lasne, Fanny Kim-Lan
    Gillet, Denis
    [J]. ADVANCES IN WEB-BASED LEARNING, ICWL 2023, 2023, 14409 : 169 - 188
  • [7] Editing Graph Visualizations by Prompting Large Language Models
    Argyriou, Evmorfia
    Boehm, Jens
    Eberle, Anne
    Gonser, Julius
    Lumpp, Anna-Lena
    Niedermann, Benjamin
    Schwarzkopf, Fabian
    [J]. GRAPH DRAWING AND NETWORK VISUALIZATION, GD 2023, PT II, 2023, 14466 : 253 - 254
  • [8] Considerations for Prompting Large Language Models-Reply
    Chen, Shan
    Savova, Guergana K.
    Bitterman, Danielle S.
    [J]. JAMA ONCOLOGY, 2024, 10 (04) : 526 - 530
  • [9] Investigating the Efficacy of Large Language Models in Reflective Assessment Methods through Chain of Thoughts Prompting
    Masikisiki, Baphumelele
    Marivate, Vukosi
    Hlophe, Yvette
    [J]. PROCEEDINGS OF THE 4TH AFRICAN CONFERENCE FOR HUMAN COMPUTER INTERACTION, AFRICHI 2023, 2023, : 44 - 49
  • [10] Guiding Large Language Models via Directional Stimulus Prompting
    Li, Zekun
    Peng, Baolin
    He, Pengcheng
    Galley, Michel
    Gao, Jianfeng
    Yan, Xifeng
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,