A Communication Theory Perspective on Prompting Engineering Methods for Large Language Models

被引:0
|
作者
Song, Yuan-Feng [1 ]
He, Yuan-Qin [1 ]
Zhao, Xue-Fang [1 ]
Gu, Han-Lin [1 ]
Jiang, Di [1 ]
Yang, Hai-Jun [1 ]
Fan, Li-Xin [1 ]
机构
[1] AI Group, WeBank Co., Ltd, Shenzhen,518000, China
关键词
Contrastive Learning - Modeling languages - Natural language processing systems - Self-supervised learning - Semi-supervised learning - Zero-shot learning;
D O I
10.1007/s11390-024-4058-8
中图分类号
学科分类号
摘要
The springing up of large language models (LLMs) has shifted the community from single-task-orientated natural language processing (NLP) research to a holistic end-to-end multi-task learning paradigm. Along this line of research endeavors in the area, LLM-based prompting methods have attracted much attention, partially due to the technological advantages brought by prompt engineering (PE) as well as the underlying NLP principles disclosed by various prompting methods. Traditional supervised learning usually requires training a model based on labeled data and then making predictions. In contrast, PE methods directly use the powerful capabilities of existing LLMs (e.g., GPT-3 and GPT-4) via composing appropriate prompts, especially under few-shot or zero-shot scenarios. Facing the abundance of studies related to the prompting and the ever-evolving nature of this field, this article aims to 1) illustrate a novel perspective to review existing PE methods within the well-established communication theory framework, 2) facilitate a better/deeper understanding of developing trends of existing PE methods used in three typical tasks, and 3) shed light on promising research directions for future PE methods. © Institute of Computing Technology, Chinese Academy of Sciences 2024.
引用
收藏
页码:984 / 1004
页数:20
相关论文
共 50 条
  • [21] Grammar Prompting for Domain-Specific Language Generation with Large Language Models
    Wang, Bailin
    Wang, Zi
    Wang, Xuezhi
    Cao, Yuan
    Saurous, Rif A.
    Kim, Yoon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [23] Do Language Models Enjoy Their Own Stories? Prompting Large Language Models for Automatic Story Evaluation
    Chhun, Cyril
    Suchanek, Fabian M.
    Clavel, Chloe
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 1122 - 1142
  • [24] Does Metacognitive Prompting Improve Causal Inference in Large Language Models?
    Ohtani, Ryusei
    Sakurai, Yuko
    Oyama, Satoshi
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 458 - 459
  • [25] On Hardware Security Bug Code Fixes by Prompting Large Language Models
    Ahmad, Baleegh
    Thakur, Shailja
    Tan, Benjamin
    Karri, Ramesh
    Pearce, Hammond
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 4043 - 4057
  • [26] Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
    Wei, Jason
    Wang, Xuezhi
    Schuurmans, Dale
    Bosma, Maarten
    Ichter, Brian
    Xia, Fei
    Chi, Ed H.
    Le, Quoc V.
    Zhou, Denny
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [27] Risk communication and large language models
    Sledge, Daniel
    Thomas, Herschel F.
    RISK HAZARDS & CRISIS IN PUBLIC POLICY, 2024,
  • [28] Ontology engineering with Large Language Models
    Mateiu, Patricia
    Groza, Adrian
    2023 25TH INTERNATIONAL SYMPOSIUM ON SYMBOLIC AND NUMERIC ALGORITHMS FOR SCIENTIFIC COMPUTING, SYNASC 2023, 2023, : 226 - 229
  • [29] A Surgical Perspective on Large Language Models
    Miller, Robert
    ANNALS OF SURGERY, 2023, 278 (02) : E211 - E213
  • [30] Legal Syllogism Prompting: Teaching Large Language Models for Legal Judgment Prediction
    Jiang, Cong
    Yang, Xiaolei
    PROCEEDINGS OF THE 19TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND LAW, ICAIL 2023, 2023, : 417 - 421