Rationality of Thought Improves Reasoning in Large Language Models

被引:0
|
作者
Gou, Tian [1 ,2 ]
Zhang, Boyao [1 ,2 ]
Sun, Zhenglie [1 ,2 ]
Wang, Jing [1 ,2 ]
Liu, Fang [1 ,2 ]
Wang, Yangang [1 ,2 ]
Wang, Jue [1 ,2 ]
机构
[1] Chinese Acad Sci, Comp Network Informat Ctr, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
基金
国家重点研发计划; 北京市自然科学基金;
关键词
Large Language Models (LLMs); Zero-Shot Reasoning; Cognitive foundations of knowledge; Rationality of Thought (RoT); Cognitive Psychology; Cognitive Bias Dataset; HEURISTICS; FALLACY;
D O I
10.1007/978-981-97-5501-1_26
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While the capabilities of large language models (LLMs) have been progressively advanced, their competence in addressing intricate reasoning tasks remains inadequate, primarily due to their insufficient cognitive capabilities. To explore the cognitive proficiency of models like GPT-4, we turn to methodologies from cognitive psychology: cognitive abilities reflect rational thinking skills, and cognitive bias tasks are often used to assess rational thinking levels. In this paper, we develop a cognitive bias dataset to measure the rational thinking and cognitive levels of LLMs. Our observations indicate that GPT-4, akin to humans, exhibits limitations in its rational thinking ability. We propose a new method, "Rationality of Thought" (RoT), to prompt LLMs into a rational thinking process during task execution. This method significantly improves the accuracy of GPT-4 on the cognitive bias task by 18.7%. Cognitive capacity is also essential for tackling complex issues, therefore, we implement RoT across various reasoning tasks. Using only a zero-shot setting, RoT outperforms inference enhancement techniques such as CoT using zero-shot, such as SVAMP(+1.8),AQUA-RAT (+6.0), ARC-c (+4.1),ARCe(+3.9) in multiple arithmetic and common sense reasoning tasks. Our empirical evaluation shows that RoT helps LLMs elevate their cognitive capabilities through rational thinking, thereby becoming more adept at navigating complex reasoning tasks.
引用
收藏
页码:343 / 358
页数:16
相关论文
共 50 条
  • [21] Large Language Models for Mathematical Reasoning: Progresses and Challenges
    Ahn, Janice
    Verma, Rishu
    Lou, Renze
    Zhang, Rui
    Yin, Wenpeng
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: STUDENT RESEARCH WORKSHOP, 2024, : 225 - 237
  • [22] The use of large language models as scaffolds for proleptic reasoning
    Olya Kudina
    Brian Ballsun-Stanton
    Mark Alfano
    Asian Journal of Philosophy, 4 (1):
  • [23] The Impact of Reasoning Step Length on Large Language Models
    Jin, Mingyu
    Yu, Qinkai
    Dong, Shu
    Zhao, Haiyan
    Hua, Wenyue
    Meng, Yanda
    Zhang, Yongfeng
    Du, Mengnan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 1830 - 1842
  • [24] TRAM: Benchmarking Temporal Reasoning for Large Language Models
    Wang, Yuqing
    Zhao, Yun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 6389 - 6415
  • [25] EconNLI: Evaluating Large Language Models on Economics Reasoning
    Guo, Yue
    Yang, Yi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 982 - 994
  • [26] Evaluating Large Language Models for Tax Law Reasoning
    Cavalcante Presa, Joao Paulo
    Camilo Junior, Celso Goncalves
    Teles de Oliveira, Savio Salvarino
    INTELLIGENT SYSTEMS, BRACIS 2024, PT I, 2025, 15412 : 460 - 474
  • [27] Targeted training for numerical reasoning with large language models
    Li, Xiao
    Liu, Sichen
    Zhu, Yin
    Cheng, Gong
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (01) : 197 - 221
  • [28] Automatic Model Selection with Large Language Models for Reasoning
    Zhao, James Xu
    Xie, Yuxi
    Kawaguchi, Kenji
    He, Junxian
    Xie, Michael Qizhe
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 758 - 783
  • [29] NEWTON: Are Large Language Models Capable of Physical Reasoning?
    Wang, Yi Ru
    Du, Jiafei
    Fox, Dieter
    Srinivasa, Siddhartha
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9743 - 9758
  • [30] Dynamic Voting for Efficient Reasoning in Large Language Models
    Xue, Mingfeng
    Liu, Dayiheng
    Lei, Wenqiang
    Ren, Xingzhang
    Yang, Baosong
    Xie, Jun
    Zhang, Yidan
    Peng, Dezhong
    Lv, Jiancheng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3085 - 3104