Can large language models effectively reason about adverse weather conditions?

被引:0
|
作者
Zafarmomen, Nima [1 ]
Samadi, Vidya [1 ,2 ]
机构
[1] Clemson Univ, Dept Agr Sci, Clemson, SC 29634 USA
[2] Clemson Univ, Artificial Intelligence Res Inst Sci & Engn AIRISE, Sch Comp, Clemson, SC USA
基金
美国国家科学基金会;
关键词
Large language model; Text classification; LLaMA; BART; BERT; Adverse weather conditions;
D O I
10.1016/j.envsoft.2025.106421
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This paper seeks to answer the question "can Large Language Models (LLMs) effectively reason about adverse weather conditions?". To address this question, we utilized multiple LLMs to harness the US National Weather Service (NWS) flood report data spanning from June 2005 to September 2024. Bidirectional and Auto-Regressive Transformer (BART), Bidirectional Encoder Representations from Transformers (BERT), Large Language Model Meta AI (LLaMA-2), LLaMA-3, and LLaMA-3.1 were employed to categorize data based on predefined labels. The methodology was implemented in Charleston County, South Carolina, USA. Extreme events were unevenly distributed across the training period with the "Cyclonic" category exhibiting significantly fewer instances compared to the "Flood" and "Thunderstorm" categories. Analysis suggests that the LLaMA-3 reached its peak performance at 60% of the dataset size while other LLMs achieved peak performance at approximately 80-100% of the dataset size. This study provided deep insights into the application of LLMs in reasoning adverse weather conditions.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Can ChatGPT Truly Overcome Other Large Language Models?
    Ray, Partha
    CANADIAN ASSOCIATION OF RADIOLOGISTS JOURNAL-JOURNAL DE L ASSOCIATION CANADIENNE DES RADIOLOGISTES, 2024, 75 (02): : 429 - 429
  • [42] BLINK: Multimodal Large Language Models Can See but Not Perceive
    Fu, Xingyu
    Hu, Yushi
    Li, Bangzheng
    Feng, Yu
    Wang, Haoyu
    Lin, Xudong
    Roth, Dan
    Smith, Noah A.
    Ma, Wei-Chiu
    Krishna, Ranjay
    COMPUTER VISION - ECCV 2024, PT XXIII, 2025, 15081 : 148 - 166
  • [43] How can large language models assist with a FRAM analysis?
    Sujan, M.
    Slater, D.
    Crumpton, E.
    SAFETY SCIENCE, 2025, 181
  • [44] Can Large Language Models Capture Dissenting Human Voices?
    Lee, Noah
    An, Na Min
    Thorne, James
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 4569 - 4585
  • [45] Can Large Language Models Better Predict Software Vulnerability?
    Katsadouros, Evangelos
    Patrikakis, Charalampos Z.
    Hurlburt, George
    IT PROFESSIONAL, 2023, 25 (03) : 4 - 8
  • [46] Can Large Language Models be Anomaly Detectors for Time Series?
    Alnegheimish, Sarah
    Nguyen, Linh
    Berti-Equille, Laure
    Veeramachaneni, Kalyan
    2024 IEEE 11TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS, DSAA 2024, 2024, : 218 - 227
  • [47] Receive, Reason, and React: Drive as You Say, With Large Language Models in Autonomous Vehicles
    Cui, Can
    Ma, Yunsheng
    Cao, Xu
    Ye, Wenqian
    Wang, Ziran
    IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2024, 16 (04) : 81 - 94
  • [48] Thinking about the language models and what we can do with them
    Horvath, Roman
    JOURNAL OF LANGUAGE AND CULTURAL EDUCATION, 2023, 11 (01) : 38 - 45
  • [49] What Can Language Models Tell Us About Human Cognition?
    Connell, Louise
    Lynott, Dermot
    CURRENT DIRECTIONS IN PSYCHOLOGICAL SCIENCE, 2024, : 181 - 189
  • [50] Can ChatGPT Detect Intent? Evaluating Large Language Models for Spoken Language Understanding
    He, Mutian
    Garner, Philip N.
    INTERSPEECH 2023, 2023, : 1109 - 1113