Probing the Consistency of Situational Information Extraction with Large Language Models: A Case Study on Crisis Computing

被引:0
|
作者
Salfinger, Andrea [1 ]
Snidaro, Lauro [1 ]
机构
[1] Univ Udine, Dept Math Comp Sci & Phys, Udine, Italy
基金
奥地利科学基金会;
关键词
Large Language Models; Crisis Management; Situation Awareness; Soft Fusion; High-Level Information Fusion;
D O I
10.1109/CogSIMA61085.2024.10553903
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recently introduced foundation models for language modeling, also known as Large Language Models (LLMs), have demonstrated breakthrough capabilities on text summarization and contextualized natural language processing. However, these also suffer from inherent deficiencies like the occasional generation of factually wrong information, known as hallucinations, and a weak consistency of produced answers strongly varying with the exact phrasing of their input query, i.e., prompt. Hence, this raises the question whether and how LLMs could replace or complement traditional information extraction and fusion modules in information fusion pipelines involving textual input sources. We empirically examine this question on a case study from crisis computing, taken from the established CrisisFacts benchmark dataset, by probing an LLM's situation understanding and summarization capabilities on the target task of extracting information relevant for establishing crisis situation awareness from social media corpora. Since social media messages are exchanged in real-time, typically targeting human readers aware of the situational context, this domain represents a prime testbed for evaluating LLMs' situational information extraction capabilities. In this work, we specifically investigate the consistency of extracted information across different model configurations and different but semantically similar prompts, which represents a crucial prerequisite for a reliable and trustworthy information extraction component.
引用
收藏
页码:91 / 98
页数:8
相关论文
共 50 条
  • [1] Probing into the Fairness of Large Language Models: A Case Study of ChatGPT
    Li, Yunqi
    Zhang, Lanjing
    Zhang, Yongfeng
    2024 58TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS, CISS, 2024,
  • [2] Large language models for generative information extraction: a survey
    Xu, Derong
    Chen, Wei
    Peng, Wenjun
    Zhang, Chao
    Xu, Tong
    Zhao, Xiangyu
    Wu, Xian
    Zheng, Yefeng
    Wang, Yang
    Chen, Enhong
    FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (06)
  • [3] Extraction of Subjective Information from Large Language Models
    Kobayashi, Atsuya
    Yamaguchi, Saneyasu
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 1612 - 1617
  • [4] LARGE LANGUAGE MODELS FOR DATA EXTRACTION IN A SYSTEMATIC REVIEW: A CASE STUDY
    Edwards, M.
    di Ruffano, L. Ferrante
    VALUE IN HEALTH, 2024, 27 (12)
  • [5] Probing for Referential Information in Language Models
    Sorodoc, Ionut-Teodor
    Gulordava, Kristina
    Boleda, Gemma
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 4177 - 4189
  • [6] Bioregulatory event extraction using large language models: a case study of rice literature
    Xinzhi Yao
    Zhihan He
    Jingbo Xia
    Genomics & Informatics, 22 (1)
  • [7] Relation extraction using large language models: a case study on acupuncture point locations
    Li, Yiming
    Peng, Xueqing
    Li, Jianfu
    Zuo, Xu
    Peng, Suyuan
    Pei, Donghong
    Tao, Cui
    Xu, Hua
    Hong, Na
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (11) : 2622 - 2631
  • [8] Structured information extraction from scientific text with large language models
    John Dagdelen
    Alexander Dunn
    Sanghoon Lee
    Nicholas Walker
    Andrew S. Rosen
    Gerbrand Ceder
    Kristin A. Persson
    Anubhav Jain
    Nature Communications, 15
  • [9] Exploring Large Language Models for Low-Resource IT Information Extraction
    Bhavya, Bhavya
    Isaza, Paulina Toro
    Deng, Yu
    Nidd, Michael
    Azad, Amar Prakash
    Shwartz, Larisa
    Zhai, ChengXiang
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 1203 - 1212
  • [10] Structured information extraction from scientific text with large language models
    Dagdelen, John
    Dunn, Alexander
    Lee, Sanghoon
    Walker, Nicholas
    Rosen, Andrew S.
    Ceder, Gerbrand
    Persson, Kristin A.
    Jain, Anubhav
    NATURE COMMUNICATIONS, 2024, 15 (01)