Exploring Equation as a Better Intermediate Meaning Representation for Numerical Reasoning of Large Language Models

被引:0
|
作者
Wang, Dingzirui [1 ]
Dou, Longxu [1 ]
Zhang, Wenbin [2 ]
Zeng, Junyu [2 ]
Che, Wanxiang [1 ]
机构
[1] Harbin Inst Technol, Harbin, Peoples R China
[2] Yunfu Technol Beijing Co Ltd, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Numerical reasoning is a vital capability for natural language processing models to understand and process numerical information in real-world scenarios. Most current methods first generate the Intermediate Meaning Representations (IMRs) of questions and then generate answers. Current SOTA methods generate programs as IMRs with large language models (LLMs). Intuitively, equations have fewer restrictions and closer semantics to the question than programs, leading to higher generation accuracy. However, current LLMs generate equations worse than programs, where we assume that the equation data is rare in pre -training data compared to programs. So in this paper, we try to use equations as IMRs to solve the numerical reasoning task by addressing two problems: (1) Theoretically, how to prove that the equation is an IMR with higher generation accuracy than programs; (2) Empirically, how to improve the generation accuracy of equations with LLMs. For the first problem, we propose and prove a proposition to theoretically compare the generation accuracy of different IMRs. For the second problem, we present a method called Boosting Numerical Reasoning by Decomposing the Generation of Equations (BRIDGE), which can improve the accuracy of LLMs in generating equations as IMRs by reducing the tendency of generating constant expressions and programs. Our method improves the performance by 2.2%, 0.9%, and 1.7% on GSM8K, SVAMP, and Algebra datasets compared to the previous state-of-the-art methods under the single reasoning path setting. Our code and prompts are available at hfips://github.com/ziruiHIT/Bridge_for_NumericaLReasoning.
引用
收藏
页码:19116 / 19125
页数:10
相关论文
共 50 条
  • [41] Reasoning with Large Language Models on Graph Tasks: The Influence of Temperature
    Wang, Yiming
    Zhang, Ziyang
    Chen, Hanwei
    Shen, Huayi
    2024 5TH INTERNATIONAL CONFERENCE ON COMPUTER ENGINEERING AND APPLICATION, ICCEA 2024, 2024, : 630 - 634
  • [42] Over-Reasoning and Redundant Calculation of Large Language Models
    Chiang, Cheng-Han
    Lee, Hung-yi
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 161 - 169
  • [43] Exploring the Capacity of Pretrained Language Models for Reasoning about Actions and Change
    He, Weinan
    Huang, Canming
    Xiao, Zhanhao
    Liu, Yongmei
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4629 - 4643
  • [44] Exploring Abductive Reasoning in Language Models for Data-to-Text Generation
    Onderkova, Kristyna
    Nickles, Matthias
    2023 31ST IRISH CONFERENCE ON ARTIFICIAL INTELLIGENCE AND COGNITIVE SCIENCE, AICS, 2023,
  • [45] Exploring the Impact of Large Language Models on Disease Diagnosis
    Almubark, Ibrahim
    IEEE ACCESS, 2025, 13 : 8225 - 8238
  • [46] Exploring Large Language Models in a Limited Resource Scenario
    Panchbhai, Anand
    Pankanti, Smarana
    2021 11TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING, DATA SCIENCE & ENGINEERING (CONFLUENCE 2021), 2021, : 147 - 152
  • [47] Exploring large language models for microstructure evolution in materials
    Satpute, Prathamesh
    Tiwari, Saurabh
    Gupta, Maneet
    Ghosh, Supriyo
    MATERIALS TODAY COMMUNICATIONS, 2024, 40
  • [48] Exploring Variability in Risk Taking With Large Language Models
    Bhatia, Sudeep
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-GENERAL, 2024, 153 (07) : 1838 - 1860
  • [49] Exploring the Potential of Large Language Models in Computational Argumentation
    Chen, Guizhen
    Cheng, Liying
    Tuan, Luu Anh
    Bing, Lidong
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 2309 - 2330
  • [50] Exploring Capabilities of Large Language Models such as ChatGPT in Radiation
    Dennstadt, Fabio
    Hastings, Janna
    Putora, Paul Martin
    Vu, Erwin
    Fischer, Galina F.
    Suveg, Krisztian
    Glatzer, Markus
    Riggenbach, Elena
    Ha, Hong-Linh
    Cihoric, Nikola
    ADVANCES IN RADIATION ONCOLOGY, 2024, 9 (03)