Are Large Language Models Capable of Causal Reasoning for Sensing Data Analysis?

被引:0
|
作者
Hu, Zhizhang [1 ]
Zhang, Yue [1 ]
Rossi, Ryan [2 ]
Yu, Tong [2 ]
Kim, Sungchul [2 ]
Pan, Shijia [1 ]
机构
[1] Univ Calif Merced, Merced, CA 95343 USA
[2] Adobe Res, San Francisco, CA 94107 USA
关键词
Large Language Model; Causal Data Reasoning;
D O I
10.1145/3662006.3662064
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The correlation analysis between socioeconomic factors and environmental impact is essential for policy making to ensure sustainability and economic development simultaneously. With the development of Internet of Things (IoT), citizen science IoT monitoring provides valuable environmental measurements, such as PM 2.5 for air quality monitoring. However, socioeconomic factors are usually interconnected and confound each other, making accurate correlation analysis challenging. To isolate this information on an individual socioeconomic factor, we need to mitigate the confounding effect (e.g., propensity score matching) of other factors on the environmental sensing data. Large language models (LLMs) have shown remarkable capabilities in data reasoning, making us wonder if they can conduct causal reasoning and answer questions like "What is the most important socioeconomic factor that impacts regional air quality?" In this paper, we present a new evaluation framework named "Order-of-Thought" based on Bloom's Taxonomy pedagogical framework to quantify the LLMs' ability for causal reasoning. We apply this evaluation framework with both natural language-based and program-based prompting strategies. Our evaluation uncovers the exceptional potentials of LLMs in causal reasoning for sensing data analysis, offering valuable insights regarding their capabilities and limitations, and providing useful directions to further achieve a higher-order thought.
引用
收藏
页码:24 / 29
页数:6
相关论文
共 50 条
  • [1] CLADDER: Assessing Causal Reasoning in Language Models
    Jin, Zhijing
    Chen, Yuen
    Leeb, Felix
    Gresele, Luigi
    Kamal, Ojasv
    Lyu, Zhiheng
    Blin, Kevin
    Gonzalez, Fernando
    Kleiman-Weiner, Max
    Sachan, Mrinmaya
    Schoelkopf, Bernhard
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [2] Large Language Models Are Reasoning Teachers
    Ho, Namgyu
    Schmid, Laura
    Yun, Se-Young
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14852 - 14882
  • [3] A Causal Framework to Quantify the Robustness of Mathematical Reasoning with Language Models
    Stolfo, Alessandro
    Jin, Zhijing
    Shridhar, Kumar
    Scholkopf, Bernhard
    Sachan, Mrinmaya
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 545 - 561
  • [4] Emergent analogical reasoning in large language models
    Taylor Webb
    Keith J. Holyoak
    Hongjing Lu
    [J]. Nature Human Behaviour, 2023, 7 : 1526 - 1541
  • [5] Conversations on reasoning: Large language models in diagnosis
    Restrepo, Daniel
    Rodman, Adam
    Abdulnour, Raja-Elie
    [J]. JOURNAL OF HOSPITAL MEDICINE, 2024, 19 (08) : 731 - 735
  • [6] Inductive reasoning in humans and large language models
    Han, Simon Jerome
    Ransom, Keith J.
    Perfors, Andrew
    Kemp, Charles
    [J]. COGNITIVE SYSTEMS RESEARCH, 2024, 83
  • [7] Large Language Models are Visual Reasoning Coordinators
    Chen, Liangyu
    Li, Bo
    Shen, Sheng
    Yang, Jingkang
    Li, Chunyuan
    Keutzer, Kurt
    Darrell, Trevor
    Liu, Ziwei
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [8] Emergent analogical reasoning in large language models
    Webb, Taylor
    Holyoak, Keith J.
    Lu, Hongjing
    [J]. NATURE HUMAN BEHAVIOUR, 2023, 7 (09) : 1526 - 1541
  • [9] Causal Dataset Discovery with Large Language Models
    Liu, Junfei
    Sun, Shaotong
    Nargesian, Fatemeh
    [J]. WORKSHOP ON HUMAN-IN-THE-LOOP DATA ANALYTICS, HILDA 2024, 2024,
  • [10] CRASS: A Novel Data Set and Benchmark to Test Counterfactual Reasoning of Large Language Models
    Frohberg, Jorg
    Binder, Frank
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 2126 - 2140