Mathematical Reasoning via Multi-step Self Questioning and Answering for Small Language Models

被引:0
|
作者
Chen, Kaiyuan [1 ]
Wang, Jin [1 ]
Zhang, Xuejie [1 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming, Yunnan, Peoples R China
基金
中国国家自然科学基金;
关键词
Mathematical Reasoning; Knowledge Distillation; Small Language Models;
D O I
10.1007/978-981-97-9440-9_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mathematical reasoning is challenging for large language models (LLMs), while the scaling relationship concerning LLM capacity is under-explored. Existing works have tried to leverage the rationales of LLMs to train small language models (SLMs) for enhanced reasoning abilities, referred to as distillation. However, most existing distillation methods have not considered guiding the small models to solve problems progressively from simple to complex, which can be a more effective way. This study proposes a multi-step self questioning and answering (M-SQA) method that guides SLMs to solve complex problems by starting from simple ones. Initially, multi-step self-questioning and answering rationales are extracted from LLMs based on complexity-based prompting. Subsequently, these rationales are employed for distilling SLMs in a multi-task learning framework, during which the model learns to multi-step reason in a self questioning and answering way and answer each sub-question in a single step iteratively. Experiments on current mathematical reasoning tasks demonstrate the effectiveness of the proposed approach.
引用
收藏
页码:81 / 93
页数:13
相关论文
共 50 条
  • [41] Using counterfactual contrast to improve compositional generalization for multi-step quantitative reasoning
    Nourbakhsh, Armineh
    Shah, Sameena
    Rose, Carolyn
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14930 - 14943
  • [42] Residual Connection-Based Multi-step Reasoning via Commonsense Knowledge for Multiple Choice Machine Reading Comprehension
    Sheng, Yixuan
    Lan, Man
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT III, 2019, 11955 : 340 - 352
  • [43] Learning multi-step prediction models for receding horizon control
    Terzi, Enrico
    Fagiano, Lorenzo
    Farina, Marcello
    Scattolini, Riccardo
    2018 EUROPEAN CONTROL CONFERENCE (ECC), 2018, : 1335 - 1340
  • [44] Multi-Step Generalized Policy Improvement by Leveraging Approximate Models
    Alegre, Lucas N.
    Bazzan, Ana L. C.
    Nowe, Ann
    da Silva, Bruno C.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [45] Multi-step estimators and shrinkage effect in time series models
    Svetunkov, Ivan
    Kourentzes, Nikolaos
    Killick, Rebecca
    COMPUTATIONAL STATISTICS, 2024, 39 (03) : 1203 - 1239
  • [46] A multi-step method for matrix condensation of finite element models
    Qu, ZQ
    JOURNAL OF SOUND AND VIBRATION, 1998, 214 (05) : 965 - 971
  • [47] Teaching Small Language Models to Reason for Knowledge-Intensive Multi-Hop Question Answering
    Li, Xiang
    He, Shizhu
    Lei, Fangyu
    Yang, Jun
    Su, Tianhuang
    Liu, Kang
    Zhao, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 7804 - 7816
  • [48] MULTI-STEP METHODS FOR MACHINE LEARNING MODELS WITH WEB METRICS
    Popchev, Ivan
    Orozova, Daniela
    COMPTES RENDUS DE L ACADEMIE BULGARE DES SCIENCES, 2023, 76 (11): : 1707 - 1715
  • [49] A note on multi-step forecasting with functional coefficient autoregressive models
    Harvill, JL
    Ray, BK
    INTERNATIONAL JOURNAL OF FORECASTING, 2005, 21 (04) : 717 - 727
  • [50] Multi-step estimators and shrinkage effect in time series models
    Ivan Svetunkov
    Nikolaos Kourentzes
    Rebecca Killick
    Computational Statistics, 2024, 39 : 1203 - 1239