Multi-party Goal Tracking with LLMs: Comparing Pre-training, Fine-tuning, and Prompt Engineering

被引:0
|
作者
Addlesee, Angus [1 ]
Sieinska, Weronika [1 ]
Gunson, Nancie [1 ]
Garcia, Daniel Hernandez [1 ]
Dondrup, Christian [1 ]
Lemon, Oliver [1 ,2 ,3 ]
机构
[1] Heriot Watt Univ, Edinburgh, Midlothian, Scotland
[2] Alana AI, London, England
[3] Edinburgh Ctr Robot, Edinburgh, Midlothian, Scotland
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper evaluates the extent to which current Large Language Models (LLMs) can capture task-oriented multi-party conversations (MPCs). We have recorded and transcribed 29 MPCs between patients, their companions, and a social robot in a hospital. We then annotated this corpus for multi-party goal-tracking and intent-slot recognition. People share goals, answer each other's goals, and provide other people's goals in MPCs - none of which occur in dyadic interactions. To understand user goals in MPCs, we compared three methods in zero-shot and few-shot settings: we fine-tuned T5, created pre-training tasks to train DialogLM using LED, and employed prompt engineering techniques with GPT-3.5-turbo, to determine which approach can complete this novel task with limited data. GPT-3.5-turbo significantly outperformed the others in a few-shot setting. The 'reasoning' style prompt, when given 7% of the corpus as example annotated conversations, was the best performing method. It correctly annotated 62.32% of the goal tracking MPCs, and 69.57% of the intent-slot recognition MPCs. A 'story' style prompt increased model hallucination, which could be detrimental if deployed in safety-critical settings. We conclude that multi-party conversations still challenge state-of-the-art LLMs.
引用
收藏
页码:229 / 241
页数:13
相关论文
共 50 条
  • [41] Evaluation of Dataset Selection for Pre-Training and Fine-Tuning Transformer Language Models for Clinical Question Answering
    Soni, Sarvesh
    Roberts, Kirk
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 5532 - 5538
  • [42] From pre-training to fine-tuning: An in-depth analysis of Large Language Models in the biomedical domain
    Bonfigli, Agnese
    Bacco, Luca
    Merone, Mario
    Dell'Orletta, Felice
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 157
  • [43] FactGen: Faithful Text Generation by Factuality-aware Pre-training and Contrastive Ranking Fine-tuning
    Lan Z.
    Li W.
    Su J.
    Xiao X.
    Liu J.
    Wu W.
    Lyu Y.
    Journal of Artificial Intelligence Research, 2023, 76 : 1281 - 1303
  • [44] FactGen: Faithful Text Generation by Factuality-aware Pre-training and Contrastive Ranking Fine-tuning
    Lan, Zhibin
    Li, Wei
    Su, Jinsong
    Xiao, Xinyan
    Liu, Jiachen
    Wu, Wenhao
    Lyu, Yajuan
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2023, 76 : 1281 - 1303
  • [45] CREATER: CTR-driven Advertising Text Generation with Controlled Pre-Training and Contrastive Fine-Tuning
    Wei, Penghui
    Yang, Xuanhua
    Liu, Shaoguo
    Wang, Liang
    Zheng, Bo
    2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2022, 2022, : 9 - 17
  • [46] Pre-training using pseudo images and fine-tuning using real images for nighttime traffic Sign Detection
    Yamamoto M.
    Ohashi G.
    IEEJ Transactions on Electronics, Information and Systems, 2021, 141 (09) : 969 - 976
  • [47] Training Deep Spiking Convolutional Neural Networks With STDP-Based Unsupervised Pre-training Followed by Supervised Fine-Tuning
    Lee, Chankyu
    Panda, Priyadarshini
    Srinivasan, Gopalakrishnan
    Roy, Kaushik
    FRONTIERS IN NEUROSCIENCE, 2018, 12
  • [48] Breaking the Barrier Between Pre-training and Fine-tuning: A Hybrid Prompting Model for Knowledge-Based VQA
    Sun, Zhongfan
    Hu, Yongli
    Gao, Qingqing
    Jiang, Huajie
    Gao, Junbin
    Sun, Yanfeng
    Yin, Baocai
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4065 - 4073
  • [49] Adaptive Pre-Training and Collaborative Fine-Tuning: A Win-Win Strategy to Improve Review Analysis Tasks
    Mao, Qianren
    Li, Jianxin
    Lin, Chenghua
    Chen, Congwen
    Peng, Hao
    Wang, Lihong
    Yu, Philip S.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 622 - 634
  • [50] Pre-training and Fine-tuning Neural Topic Model: A Simple yet Effective Approach to Incorporating External Knowledge
    Zhang, Linhai
    Hu, Xumeng
    Wang, Boyu
    Zhou, Deyu
    Zhang, Qian-Wen
    Cao, Yunbo
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5980 - 5989