Targeted training for numerical reasoning with large language models

被引:0
|
作者
Li, Xiao [1 ]
Liu, Sichen [1 ]
Zhu, Yin [1 ]
Cheng, Gong [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
基金
中国国家自然科学基金;
关键词
Data refining; Numerical reasoning; Large language model; Knowledge distillation;
D O I
10.1007/s10115-024-02216-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
After recent gains achieved by large language models (LLMs) on numerical reasoning tasks, it has become of interest to have LLMs teach small models to improve on numerical reasoning. Instructing LLMs to generate Chains of Thought to fine-tune small models is an established approach. However, small models are passive in this line of work and may not be able to exploit the provided training data. In this paper, we propose a novel targeted training strategy to match LLM's assistance with small models' capacities. The small model will proactively request LLM's assistance when it sifts out confusing training data. Then, LLM refines such data by successively revising reasoning steps and reducing question complexity before feeding the small model. Experiments show that this targeted training approach remarkably improves the performance of small models on a range of numerical reasoning datasets by 12-25%, making small models even competitive with some LLMs.
引用
收藏
页数:25
相关论文
共 50 条
  • [1] Large Language Models Are Reasoning Teachers
    Ho, Namgyu
    Schmid, Laura
    Yun, Se-Young
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14852 - 14882
  • [2] Injecting Numerical Reasoning Skills into Language Models
    Geva, Mor
    Gupta, Ankit
    Beran, Jonathan
    [J]. 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 946 - 958
  • [3] Exploring Equation as a Better Intermediate Meaning Representation for Numerical Reasoning of Large Language Models
    Wang, Dingzirui
    Dou, Longxu
    Zhang, Wenbin
    Zeng, Junyu
    Che, Wanxiang
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19116 - 19125
  • [4] Emergent analogical reasoning in large language models
    Taylor Webb
    Keith J. Holyoak
    Hongjing Lu
    [J]. Nature Human Behaviour, 2023, 7 : 1526 - 1541
  • [5] Conversations on reasoning: Large language models in diagnosis
    Restrepo, Daniel
    Rodman, Adam
    Abdulnour, Raja-Elie
    [J]. JOURNAL OF HOSPITAL MEDICINE, 2024, 19 (08) : 731 - 735
  • [6] Inductive reasoning in humans and large language models
    Han, Simon Jerome
    Ransom, Keith J.
    Perfors, Andrew
    Kemp, Charles
    [J]. COGNITIVE SYSTEMS RESEARCH, 2024, 83
  • [7] Large Language Models are Visual Reasoning Coordinators
    Chen, Liangyu
    Li, Bo
    Shen, Sheng
    Yang, Jingkang
    Li, Chunyuan
    Keutzer, Kurt
    Darrell, Trevor
    Liu, Ziwei
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [8] Emergent analogical reasoning in large language models
    Webb, Taylor
    Holyoak, Keith J.
    Lu, Hongjing
    [J]. NATURE HUMAN BEHAVIOUR, 2023, 7 (09) : 1526 - 1541
  • [9] Reasoning with large language models for medical question answering
    Lucas, Mary M.
    Yang, Justin
    Pomeroy, Jon K.
    Yang, Christopher C.
    [J]. JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09)
  • [10] Rationality of Thought Improves Reasoning in Large Language Models
    Gou, Tian
    Zhang, Boyao
    Sun, Zhenglie
    Wang, Jing
    Liu, Fang
    Wang, Yangang
    Wang, Jue
    [J]. KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2024, 2024, 14887 : 343 - 358