Probing the Consistency of Situational Information Extraction with Large Language Models: A Case Study on Crisis Computing

被引:0
|
作者
Salfinger, Andrea [1 ]
Snidaro, Lauro [1 ]
机构
[1] Univ Udine, Dept Math Comp Sci & Phys, Udine, Italy
基金
奥地利科学基金会;
关键词
Large Language Models; Crisis Management; Situation Awareness; Soft Fusion; High-Level Information Fusion;
D O I
10.1109/CogSIMA61085.2024.10553903
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recently introduced foundation models for language modeling, also known as Large Language Models (LLMs), have demonstrated breakthrough capabilities on text summarization and contextualized natural language processing. However, these also suffer from inherent deficiencies like the occasional generation of factually wrong information, known as hallucinations, and a weak consistency of produced answers strongly varying with the exact phrasing of their input query, i.e., prompt. Hence, this raises the question whether and how LLMs could replace or complement traditional information extraction and fusion modules in information fusion pipelines involving textual input sources. We empirically examine this question on a case study from crisis computing, taken from the established CrisisFacts benchmark dataset, by probing an LLM's situation understanding and summarization capabilities on the target task of extracting information relevant for establishing crisis situation awareness from social media corpora. Since social media messages are exchanged in real-time, typically targeting human readers aware of the situational context, this domain represents a prime testbed for evaluating LLMs' situational information extraction capabilities. In this work, we specifically investigate the consistency of extracted information across different model configurations and different but semantically similar prompts, which represents a crucial prerequisite for a reliable and trustworthy information extraction component.
引用
收藏
页码:91 / 98
页数:8
相关论文
共 50 条
  • [31] Crack image classification and information extraction in steel bridges using multimodal large language models
    Wang, Xiao
    Yue, Qingrui
    Liu, Xiaogang
    AUTOMATION IN CONSTRUCTION, 2025, 171
  • [32] Leveraging Medical Knowledge Graphs and Large Language Models for Enhanced Mental Disorder Information Extraction
    Park, Chaelim
    Lee, Hayoung
    Jeong, Ok-ran
    FUTURE INTERNET, 2024, 16 (08)
  • [33] Revisiting Relation Extraction in the era of Large Language Models
    Wadhwa, Somin
    Amir, Silvio
    Wallace, Byron C.
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15566 - 15589
  • [34] Trend Extraction and Analysis via Large Language Models
    Soru, Tommaso
    Marshall, Jim
    18TH IEEE INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING, ICSC 2024, 2024, : 285 - 288
  • [35] Large Language Models as Evaluators in Education: Verification of Feedback Consistency and Accuracy
    Seo, Hyein
    Hwang, Taewook
    Jung, Jeesu
    Kang, Hyeonseok
    Namgoong, Hyuk
    Lee, Yohan
    Jung, Sangkeun
    APPLIED SCIENCES-BASEL, 2025, 15 (02):
  • [36] Evaluating the Factual Consistency of Large Language Models Through News Summarization
    Tam, Derek
    Mascarenhas, Anisha
    Zhang, Shiyue
    Kwan, Sarah
    Bansal, Mohit
    Raffel, Colin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5220 - 5255
  • [37] Lexical Semantics with Large Language Models: A Case Study of English break
    Petersen, Erika
    Potts, Christopher
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 490 - 511
  • [38] Probing the "Creativity" of Large Language Models: Can Models Produce Divergent Semantic Association?
    Chen, Honghua
    Ding, Nai
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12881 - 12888
  • [39] Probing Toxic Content in Large Pre-Trained Language Models
    Ousidhoum, Nedjma
    Zhao, Xinran
    Fang, Tianqing
    Song, Yangqiu
    Yeung, Dit-Yan
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4262 - 4274
  • [40] Ethics, Governance, and User Mental Models for Large Language Models in Computing Education
    Zhou, Kyrie Zhixuan
    Kilhoffer, Zachary
    Sanfilippo, Madelyn Rose
    Underwood, Ted
    Gumusel, Ece
    Wei, Mengyi
    Choudhry, Abhinav
    Xiong, Jinjun
    XRDS: Crossroads, 2024, 31 (01): : 46 - 51