Rationality of Thought Improves Reasoning in Large Language Models

被引:0
|
作者
Gou, Tian [1 ,2 ]
Zhang, Boyao [1 ,2 ]
Sun, Zhenglie [1 ,2 ]
Wang, Jing [1 ,2 ]
Liu, Fang [1 ,2 ]
Wang, Yangang [1 ,2 ]
Wang, Jue [1 ,2 ]
机构
[1] Chinese Acad Sci, Comp Network Informat Ctr, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
基金
国家重点研发计划; 北京市自然科学基金;
关键词
Large Language Models (LLMs); Zero-Shot Reasoning; Cognitive foundations of knowledge; Rationality of Thought (RoT); Cognitive Psychology; Cognitive Bias Dataset; HEURISTICS; FALLACY;
D O I
10.1007/978-981-97-5501-1_26
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While the capabilities of large language models (LLMs) have been progressively advanced, their competence in addressing intricate reasoning tasks remains inadequate, primarily due to their insufficient cognitive capabilities. To explore the cognitive proficiency of models like GPT-4, we turn to methodologies from cognitive psychology: cognitive abilities reflect rational thinking skills, and cognitive bias tasks are often used to assess rational thinking levels. In this paper, we develop a cognitive bias dataset to measure the rational thinking and cognitive levels of LLMs. Our observations indicate that GPT-4, akin to humans, exhibits limitations in its rational thinking ability. We propose a new method, "Rationality of Thought" (RoT), to prompt LLMs into a rational thinking process during task execution. This method significantly improves the accuracy of GPT-4 on the cognitive bias task by 18.7%. Cognitive capacity is also essential for tackling complex issues, therefore, we implement RoT across various reasoning tasks. Using only a zero-shot setting, RoT outperforms inference enhancement techniques such as CoT using zero-shot, such as SVAMP(+1.8),AQUA-RAT (+6.0), ARC-c (+4.1),ARCe(+3.9) in multiple arithmetic and common sense reasoning tasks. Our empirical evaluation shows that RoT helps LLMs elevate their cognitive capabilities through rational thinking, thereby becoming more adept at navigating complex reasoning tasks.
引用
收藏
页码:343 / 358
页数:16
相关论文
共 50 条
  • [41] Exploring Reversal Mathematical Reasoning Ability for Large Language Models
    Guo, Pei
    You, Wangjie
    Li, Juntao
    Yan, Bowen
    Zhang, Min
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 13671 - 13685
  • [42] LARGE MARGIN TRAINING IMPROVES LANGUAGE MODELS FOR ASR
    Wang, Jilin
    Huang, Jiaji
    Church, Kenneth Ward
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7368 - 7372
  • [43] Large Language Models With Holistically Thought Could Be Better Doctors
    Weng, Yixuan
    Li, Bin
    Xia, Fei
    Zhu, Minjun
    Sun, Bin
    He, Shizhu
    Liu, Shengping
    Li, Kang
    Li, Shutao
    Zhao, Jun
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT II, NLPCC 2024, 2025, 15360 : 319 - 332
  • [44] Active Prompting with Chain-of-Thought for Large Language Models
    Diao, Shizhe
    Wang, Pengcheng
    Lin, Yong
    Pan, Rui
    Liu, Xiang
    Zhang, Tong
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 1330 - 1350
  • [45] Chain of Thought Utilization in Large Language Models and Application in Nephrology
    Miao, Jing
    Thongprayoon, Charat
    Suppadungsuk, Supawadee
    Krisanapan, Pajaree
    Radhakrishnan, Yeshwanter
    Cheungpasitporn, Wisit
    MEDICINA-LITHUANIA, 2024, 60 (01):
  • [46] Self-prompted Chain-of-Thought on Large Language Models for Open-domain Multi-hop Reasoning
    Wang, Jinyuan
    Li, Junlong
    Zhao, Hai
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 2717 - 2731
  • [47] Masked Thought: Simply Masking Partial Reasoning Steps Can Improve Mathematical Reasoning Learning of Language Models
    Chen, Changyu
    Wang, Xiting
    Lin, Ting-En
    Lv, Ang
    Wu, Yuchuan
    Gao, Xin
    Wen, Ji-Rong
    Yan, Rui
    Li, Yongbin
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5872 - 5900
  • [48] Understanding Social Reasoning in Language Models with Language Models
    Gandhi, Kanishk
    Franken, J. -Philipp
    Gerstenberg, Tobias
    Goodman, Noah D.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [49] Towards Benchmarking and Improving the Temporal Reasoning Capability of Large Language Models
    Tan, Qingyu
    Ng, Hwee Tou
    Bing, Lidong
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14820 - 14835
  • [50] An Evaluation of Reasoning Capabilities of Large Language Models in Financial Sentiment Analysis
    Du, Kelvin
    Xing, Frank
    Mao, Rui
    Cambria, Erik
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 189 - 194