Mathematical Reasoning via Multi-step Self Questioning and Answering for Small Language Models

被引:0
|
作者
Chen, Kaiyuan [1 ]
Wang, Jin [1 ]
Zhang, Xuejie [1 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming, Yunnan, Peoples R China
基金
中国国家自然科学基金;
关键词
Mathematical Reasoning; Knowledge Distillation; Small Language Models;
D O I
10.1007/978-981-97-9440-9_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mathematical reasoning is challenging for large language models (LLMs), while the scaling relationship concerning LLM capacity is under-explored. Existing works have tried to leverage the rationales of LLMs to train small language models (SLMs) for enhanced reasoning abilities, referred to as distillation. However, most existing distillation methods have not considered guiding the small models to solve problems progressively from simple to complex, which can be a more effective way. This study proposes a multi-step self questioning and answering (M-SQA) method that guides SLMs to solve complex problems by starting from simple ones. Initially, multi-step self-questioning and answering rationales are extracted from LLMs based on complexity-based prompting. Subsequently, these rationales are employed for distilling SLMs in a multi-task learning framework, during which the model learns to multi-step reason in a self questioning and answering way and answer each sub-question in a single step iteratively. Experiments on current mathematical reasoning tasks demonstrate the effectiveness of the proposed approach.
引用
收藏
页码:81 / 93
页数:13
相关论文
共 50 条
  • [21] Small Language Models Need Strong Verifiers to Self-Correct Reasoning
    Zhang, Yunxiang
    Khalifa, Muhammad
    Logeswaran, Lajanugen
    Kim, Jaekyeom
    Lee, Moontae
    Lee, Honglak
    Wang, Lu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 15637 - 15653
  • [22] Understanding world models through multi-step pruning policy via reinforcement learning
    He, Zhiqiang
    Qiu, Wen
    Zhao, Wei
    Shao, Xun
    Liu, Zhi
    INFORMATION SCIENCES, 2025, 686
  • [23] Large Language Models for Mathematical Reasoning: Progresses and Challenges
    Ahn, Janice
    Verma, Rishu
    Lou, Renze
    Zhang, Rui
    Yin, Wenpeng
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: STUDENT RESEARCH WORKSHOP, 2024, : 225 - 237
  • [24] Automated Assessment of Multi-Step Answers for Mathematical Word Problems
    Kadupitiya, J. C. S.
    Ranathunga, Surangika
    Dias, Gihan
    2016 SIXTEENTH INTERNATIONAL CONFERENCE ON ADVANCES IN ICT FOR EMERGING REGIONS (ICTER) - 2016, 2016, : 66 - 71
  • [25] Multi-step Forecasting via Multi-task Learning
    Jawed, Shayan
    Rashed, Ahmed
    Schmidt-Thieme, Lars
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 790 - 799
  • [26] Multi-Step Reasoning Over Unstructured Text with Beam Dense Retrieval
    Zhao, Chen
    Xiong, Chenyan
    Boyd-Graber, Jordan
    Daume, Hal, III
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 4635 - 4641
  • [27] Probing Cross-Modal Representations in Multi-Step Relational Reasoning
    Parfenova, Iuliia
    Elliott, Desmond
    Fernandez, Raquel
    Pezzelle, Sandro
    REPL4NLP 2021: PROCEEDINGS OF THE 6TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP, 2021, : 152 - 162
  • [28] A Mechanistic Analysis of a Transformer Trained on a Symbolic Multi-Step Reasoning Task
    Brinkmann, Jannik
    Sheshadri, Abhay
    Levoso, Victor
    Swoboda, Paul
    Bartelt, Christian
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 4082 - 4102
  • [29] The Impact of Reasoning Step Length on Large Language Models
    Jin, Mingyu
    Yu, Qinkai
    Dong, Shu
    Zhao, Haiyan
    Hua, Wenyue
    Meng, Yanda
    Zhang, Yongfeng
    Du, Mengnan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 1830 - 1842
  • [30] Triggering Multi-Hop Reasoning for Question Answering in Language Models using Soft Prompts and Random Walks
    Misra, Kanishka
    dos Santos, Cicero Nogueira
    Shakeri, Siamak
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 972 - 985