Exploring Equation as a Better Intermediate Meaning Representation for Numerical Reasoning of Large Language Models

被引:0
|
作者
Wang, Dingzirui [1 ]
Dou, Longxu [1 ]
Zhang, Wenbin [2 ]
Zeng, Junyu [2 ]
Che, Wanxiang [1 ]
机构
[1] Harbin Inst Technol, Harbin, Peoples R China
[2] Yunfu Technol Beijing Co Ltd, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Numerical reasoning is a vital capability for natural language processing models to understand and process numerical information in real-world scenarios. Most current methods first generate the Intermediate Meaning Representations (IMRs) of questions and then generate answers. Current SOTA methods generate programs as IMRs with large language models (LLMs). Intuitively, equations have fewer restrictions and closer semantics to the question than programs, leading to higher generation accuracy. However, current LLMs generate equations worse than programs, where we assume that the equation data is rare in pre -training data compared to programs. So in this paper, we try to use equations as IMRs to solve the numerical reasoning task by addressing two problems: (1) Theoretically, how to prove that the equation is an IMR with higher generation accuracy than programs; (2) Empirically, how to improve the generation accuracy of equations with LLMs. For the first problem, we propose and prove a proposition to theoretically compare the generation accuracy of different IMRs. For the second problem, we present a method called Boosting Numerical Reasoning by Decomposing the Generation of Equations (BRIDGE), which can improve the accuracy of LLMs in generating equations as IMRs by reducing the tendency of generating constant expressions and programs. Our method improves the performance by 2.2%, 0.9%, and 1.7% on GSM8K, SVAMP, and Algebra datasets compared to the previous state-of-the-art methods under the single reasoning path setting. Our code and prompts are available at hfips://github.com/ziruiHIT/Bridge_for_NumericaLReasoning.
引用
收藏
页码:19116 / 19125
页数:10
相关论文
共 50 条
  • [1] Targeted training for numerical reasoning with large language models
    Li, Xiao
    Liu, Sichen
    Zhu, Yin
    Cheng, Gong
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, 67 (01) : 197 - 221
  • [2] Exploring Reversal Mathematical Reasoning Ability for Large Language Models
    Guo, Pei
    You, Wangjie
    Li, Juntao
    Yan, Bowen
    Zhang, Min
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 13671 - 13685
  • [3] Making Large Language Models Better Planners with Reasoning-Decision Alignment
    Huang, Zhijian
    Tang, Tao
    Chen, Shaoxiang
    Lin, Sihao
    Jie, Zequn
    Ma, Lin
    Wang, Guangrun
    Liang, Xiaodan
    COMPUTER VISION - ECCV 2024, PT XXXVI, 2025, 15094 : 73 - 90
  • [4] Exploring the Numerical Reasoning Capabilities of Language Models: A Comprehensive Analysis on Tabular Data
    Akhtar, Mubashara
    Shankarampeta, Abhilash
    Gupta, Vivek
    Patil, Arpit
    Cocarascul, Oana
    Simper, Elena
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 15391 - 15405
  • [5] Large Language Models Are Reasoning Teachers
    Ho, Namgyu
    Schmid, Laura
    Yun, Se-Young
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14852 - 14882
  • [6] Meaning and understanding in large language models
    Havlik, Vladimir
    SYNTHESE, 2024, 205 (01)
  • [7] Exploring the representation of Chinese cultural symbols dissemination in the era of large language models
    Zhang, Yixiao
    He, Yuan
    Xia, Yining
    Wang, Yanbo
    Dong, Xianghui
    Yao, Junchen
    INTERNATIONAL COMMUNICATION OF CHINESE CULTURE, 2024, 11 (02) : 215 - 237
  • [8] Injecting Numerical Reasoning Skills into Language Models
    Geva, Mor
    Gupta, Ankit
    Beran, Jonathan
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 946 - 958
  • [9] Towards Reasoning in Large Language Models: A Survey
    Huang, Jie
    Chang, Kevin Chen-Chuan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1049 - 1065
  • [10] Conversations on reasoning: Large language models in diagnosis
    Restrepo, Daniel
    Rodman, Adam
    Abdulnour, Raja-Elie
    JOURNAL OF HOSPITAL MEDICINE, 2024, 19 (08) : 731 - 735