Over-Reasoning and Redundant Calculation of Large Language Models

被引:0
|
作者
Chiang, Cheng-Han [1 ]
Lee, Hung-yi [1 ]
机构
[1] Natl Taiwan Univ, Taipei, Taiwan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) can solve problems step-by-step. While this chain-of-thought (CoT) reasoning boosts LLMs' performance, it is unclear if LLMs know when to use CoT and whether those CoT are always necessary to answer the question. This paper shows that LLMs tend to generate redundant calculations and reasoning on a manually constructed math QA dataset, GSM8K-Zero. GSM8K-Zero is constructed such that the questions can be answered without any calculations, but LLMs, including Llama-2 models and Claude-2, tend to generate lengthy and unnecessary calculations to answer the questions. We also conduct experiments to explain why LLMs generate redundant calculations and reasonings. GSM8K-Zero is publicly available at https://github.com/d223302/Over-Reasoning-of- LLMs and https://huggingface.co/datasets/dcml0714/GSM8K-Zero.
引用
收藏
页码:161 / 169
页数:9
相关论文
共 50 条
  • [41] Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models
    Lu, Pan
    Peng, Baolin
    Cheng, Hao
    Galley, Michel
    Chang, Kai-Wei
    Wu, Ying Nian
    Zhu, Song-Chun
    Gao, Jianfeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [42] Commonsense Reasoning and Explainable Artificial Intelligence Using Large Language Models
    Krause, Stefanie
    Stolzenburg, Frieder
    ARTIFICIAL INTELLIGENCE-ECAI 2023 INTERNATIONAL WORKSHOPS, PT 1, XAI3, TACTIFUL, XI-ML, SEDAMI, RAAIT, AI4S, HYDRA, AI4AI, 2023, 2024, 1947 : 302 - 319
  • [43] Leveraging the Inductive Bias of Large Language Models for Abstract Textual Reasoning
    Rytting, Christopher Michael
    Wingate, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [44] Can Euler Diagrams Improve Syllogistic Reasoning in Large Language Models?
    Ando, Risako
    Ozeki, Kentaro
    Morishita, Takanobu
    Abe, Hirohiko
    Mineshima, Koji
    Okada, Mitsuhiro
    DIAGRAMMATIC REPRESENTATION AND INFERENCE, DIAGRAMS 2024, 2024, 14981 : 232 - 248
  • [45] Ellipsis-Dependent Reasoning: a New Challenge for Large Language Models
    Hardt, Daniel
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 39 - 47
  • [46] Chain of Logic: Rule-Based Reasoning with Large Language Models
    Servantez, Sergio
    Barrow, Joe
    Hammond, Kristian
    Jain, Rajiv
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 2721 - 2733
  • [47] Are Large Language Models Capable of Causal Reasoning for Sensing Data Analysis?
    Hu, Zhizhang
    Zhang, Yue
    Rossi, Ryan
    Yu, Tong
    Kim, Sungchul
    Pan, Shijia
    PROCEEDINGS OF THE 2024 WORKSHOP ON EDGE AND MOBILE FOUNDATION MODELS, EDGEFM 2024, 2024, : 24 - 29
  • [48] Follow the Rules: Reasoning for Video Anomaly Detection with Large Language Models
    Yang, Yuchen
    Lee, Kwonjoon
    Dariush, Behzad
    Cao, Yinzhi
    Lo, Shao-Yuan
    COMPUTER VISION - ECCV 2024, PT LXXXI, 2025, 15139 : 304 - 322
  • [49] Multi-Agent Reasoning with Large Language Models for Effective Corporate Planning
    Tsao, Wen-Kwang
    2023 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE, CSCI 2023, 2023, : 365 - 370
  • [50] Online tools help large language models to solve problems through reasoning
    Aleksandra Piktus
    Nature, 2023, 618 : 465 - 466