Towards Analysis and Interpretation of Large Language Models for Arithmetic Reasoning

被引:0
|
作者
Akter, Mst Shapna [1 ]
Shahriar, Hossain [2 ]
Cuzzocrea, Alfredo [3 ,4 ]
机构
[1] Univ West Florida, Dept Intelligent Syst & Robot, Pensacola, FL 32514 USA
[2] Univ West Florida, Ctr Cybersecur, Pensacola, FL USA
[3] Univ Calabria, iDEA Lab, Arcavacata Di Rende, Italy
[4] Univ Paris City, Dept Comp Sci, Paris, France
关键词
LLMs; Arithmetic Reasoning; Causal Mediation Analysis;
D O I
10.1109/SDS60720.2024.00049
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs) have recently conquered the research scene, with particular regards to the Transformer architecture in the context of arithmetic reasoning. In this so-delineated scenario, this paper puts the basis for a causal mediation analysis about the approach of Transformer-based LLMs to complex arithmetic problems. In particular, we try to discover which parameters are crucial for complex reasoning tasks such as model activations. Our preliminary results state that, for complex arithmetic operations, information is channeled from mid-layer activations to the final token through enhanced attention mechanisms. Preliminary experiments are reported.
引用
收藏
页码:267 / 270
页数:4
相关论文
共 50 条
  • [21] Large Language Models for Mathematical Reasoning: Progresses and Challenges
    Ahn, Janice
    Verma, Rishu
    Lou, Renze
    Zhang, Rui
    Yin, Wenpeng
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: STUDENT RESEARCH WORKSHOP, 2024, : 225 - 237
  • [22] The use of large language models as scaffolds for proleptic reasoning
    Olya Kudina
    Brian Ballsun-Stanton
    Mark Alfano
    Asian Journal of Philosophy, 4 (1):
  • [23] The Impact of Reasoning Step Length on Large Language Models
    Jin, Mingyu
    Yu, Qinkai
    Dong, Shu
    Zhao, Haiyan
    Hua, Wenyue
    Meng, Yanda
    Zhang, Yongfeng
    Du, Mengnan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 1830 - 1842
  • [24] TRAM: Benchmarking Temporal Reasoning for Large Language Models
    Wang, Yuqing
    Zhao, Yun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 6389 - 6415
  • [25] EconNLI: Evaluating Large Language Models on Economics Reasoning
    Guo, Yue
    Yang, Yi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 982 - 994
  • [26] Evaluating Large Language Models for Tax Law Reasoning
    Cavalcante Presa, Joao Paulo
    Camilo Junior, Celso Goncalves
    Teles de Oliveira, Savio Salvarino
    INTELLIGENT SYSTEMS, BRACIS 2024, PT I, 2025, 15412 : 460 - 474
  • [27] Targeted training for numerical reasoning with large language models
    Li, Xiao
    Liu, Sichen
    Zhu, Yin
    Cheng, Gong
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (01) : 197 - 221
  • [28] Automatic Model Selection with Large Language Models for Reasoning
    Zhao, James Xu
    Xie, Yuxi
    Kawaguchi, Kenji
    He, Junxian
    Xie, Michael Qizhe
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 758 - 783
  • [29] NEWTON: Are Large Language Models Capable of Physical Reasoning?
    Wang, Yi Ru
    Du, Jiafei
    Fox, Dieter
    Srinivasa, Siddhartha
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9743 - 9758
  • [30] Dynamic Voting for Efficient Reasoning in Large Language Models
    Xue, Mingfeng
    Liu, Dayiheng
    Lei, Wenqiang
    Ren, Xingzhang
    Yang, Baosong
    Xie, Jun
    Zhang, Yidan
    Peng, Dezhong
    Lv, Jiancheng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3085 - 3104