Probing the Consistency of Situational Information Extraction with Large Language Models: A Case Study on Crisis Computing

被引:0
|
作者
Salfinger, Andrea [1 ]
Snidaro, Lauro [1 ]
机构
[1] Univ Udine, Dept Math Comp Sci & Phys, Udine, Italy
基金
奥地利科学基金会;
关键词
Large Language Models; Crisis Management; Situation Awareness; Soft Fusion; High-Level Information Fusion;
D O I
10.1109/CogSIMA61085.2024.10553903
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recently introduced foundation models for language modeling, also known as Large Language Models (LLMs), have demonstrated breakthrough capabilities on text summarization and contextualized natural language processing. However, these also suffer from inherent deficiencies like the occasional generation of factually wrong information, known as hallucinations, and a weak consistency of produced answers strongly varying with the exact phrasing of their input query, i.e., prompt. Hence, this raises the question whether and how LLMs could replace or complement traditional information extraction and fusion modules in information fusion pipelines involving textual input sources. We empirically examine this question on a case study from crisis computing, taken from the established CrisisFacts benchmark dataset, by probing an LLM's situation understanding and summarization capabilities on the target task of extracting information relevant for establishing crisis situation awareness from social media corpora. Since social media messages are exchanged in real-time, typically targeting human readers aware of the situational context, this domain represents a prime testbed for evaluating LLMs' situational information extraction capabilities. In this work, we specifically investigate the consistency of extracted information across different model configurations and different but semantically similar prompts, which represents a crucial prerequisite for a reliable and trustworthy information extraction component.
引用
收藏
页码:91 / 98
页数:8
相关论文
共 50 条
  • [21] Towards normalized clinical information extraction in Chinese radiology report with large language models
    Xu, Qinwei
    Xu, Xingkun
    Zhou, Chenyi
    Liu, Zuozhu
    Huang, Feiyue
    Li, Shaoxin
    Zhu, Lifeng
    Bai, Zhian
    Xu, Yuchen
    Hu, Weiguo
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 271
  • [22] Large language models can outperform humans in social situational judgments
    Mittelstaedt, Justin M.
    Maier, Julia
    Goerke, Panja
    Zinn, Frank
    Hermes, Michael
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [23] TrueTeacher: Learning Factual Consistency Evaluation with Large Language Models
    Gekhman, Zorik
    Herzig, Jonathan
    Aharoni, Roee
    Elkind, Chen
    Szpektor, Idan
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 2053 - 2070
  • [24] Factual consistency evaluation of summarization in the Era of large language models
    Luo, Zheheng
    Xie, Qianqian
    Ananiadou, Sophia
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 254
  • [25] Scaling and Adapting Large Language Models for Portuguese Open Information Extraction: A Comparative Study of Fine-Tuning and LoRA
    Melo, Alan
    Cabral, Bruno
    Claro, Daniela Barreiro
    INTELLIGENT SYSTEMS, BRACIS 2024, PT III, 2025, 15414 : 427 - 441
  • [26] LARGE LANGUAGE MODELS FOR RISK OF BIAS ASSESSMENT: A CASE STUDY
    Edwards, M.
    Bishop, E.
    Reddish, K.
    Carr, E.
    di Ruffano, L. Ferrante
    VALUE IN HEALTH, 2024, 27 (12)
  • [27] On Computing Paradigms - Where Will Large Language Models Be Going
    Wu, Xindong
    Zhu, Xingquan
    Baralis, Elena
    Lu, Ruqian
    Kumar, Vipin
    Rutkowski, Leszck
    Tang, Jie
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 1577 - 1582
  • [28] AI Computing Systems for Large Language Models Training
    Zhang, Zhen-Xing
    Wen, Yuan-Bo
    Lyu, Han-Qi
    Liu, Chang
    Zhang, Rui
    Li, Xia-Qing
    Wang, Chao
    Du, Zi-Dong
    Guo, Qi
    Li, Ling
    Zhou, Xue-Hai
    Chen, Yun-Ji
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2025, 40 (01) : 6 - 41
  • [29] Computing Architecture for Large-Language Models (LLMs) and Large Multimodal Models (LMMs)
    Liang, Bor-Sung
    PROCEEDINGS OF THE 2024 INTERNATIONAL SYMPOSIUM ON PHYSICAL DESIGN, ISPD 2024, 2024, : 233 - 234
  • [30] Automatic bridge inspection database construction through hybrid information extraction and large language models
    Zhang, Chenhong
    Lei, Xiaoming
    Xia, Ye
    Sun, Limin
    DEVELOPMENTS IN THE BUILT ENVIRONMENT, 2024, 20