Mathematical Reasoning via Multi-step Self Questioning and Answering for Small Language Models

被引:0
|
作者
Chen, Kaiyuan [1 ]
Wang, Jin [1 ]
Zhang, Xuejie [1 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming, Yunnan, Peoples R China
基金
中国国家自然科学基金;
关键词
Mathematical Reasoning; Knowledge Distillation; Small Language Models;
D O I
10.1007/978-981-97-9440-9_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mathematical reasoning is challenging for large language models (LLMs), while the scaling relationship concerning LLM capacity is under-explored. Existing works have tried to leverage the rationales of LLMs to train small language models (SLMs) for enhanced reasoning abilities, referred to as distillation. However, most existing distillation methods have not considered guiding the small models to solve problems progressively from simple to complex, which can be a more effective way. This study proposes a multi-step self questioning and answering (M-SQA) method that guides SLMs to solve complex problems by starting from simple ones. Initially, multi-step self-questioning and answering rationales are extracted from LLMs based on complexity-based prompting. Subsequently, these rationales are employed for distilling SLMs in a multi-task learning framework, during which the model learns to multi-step reason in a self questioning and answering way and answer each sub-question in a single step iteratively. Experiments on current mathematical reasoning tasks demonstrate the effectiveness of the proposed approach.
引用
收藏
页码:81 / 93
页数:13
相关论文
共 50 条
  • [1] Explore Multi-Step Reasoning in Video Question Answering
    Han, Yahong
    PROCEEDINGS OF THE 1ST WORKSHOP AND CHALLENGE ON COMPREHENSIVE VIDEO UNDERSTANDING IN THE WILD (COVIEW'18), 2018, : 5 - 5
  • [2] Explore Multi-Step Reasoning in Video Question Answering
    Song, Xiaomeng
    Shi, Yucheng
    Chen, Xin
    Han, Yahong
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 239 - 247
  • [3] Towards a Mechanistic Interpretation of Multi-Step Reasoning Capabilities of Language Models
    Hou, Yifan
    Li, Jiaoda
    Fei, Yu
    Stolfo, Alessandro
    Zhou, Wangchunshu
    Zeng, Guangtao
    Bosselut, Antoine
    Sachan, Mrinmaya
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 4902 - 4919
  • [4] INFORM : Information eNtropy based multi-step reasoning FOR large language Models
    Zhou, Chuyue
    You, Wangjie
    Li, Juntao
    Ye, Jing
    Chen, Kehai
    Zhang, Min
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 3565 - 3576
  • [5] MindMap: Constructing Evidence Chains for Multi-Step Reasoning in Large Language Models
    Wu, Yangyu
    Han, Xu
    Song, Wei
    Cheng, Miaomiao
    Li, Fei
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19270 - 19278
  • [6] Efficient Multi-step Reasoning Attention Network for Visual Question Answering
    Zhang, Haotian
    Wu, Wei
    Zhang, Meng
    THIRTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING (ICGIP 2021), 2022, 12083
  • [7] Neural Multi-step Reasoning for Question Answering on Semi-structured Tables
    Haug, Till
    Ganea, Octavian-Eugen
    Grnarova, Paulina
    ADVANCES IN INFORMATION RETRIEVAL (ECIR 2018), 2018, 10772 : 611 - 617
  • [8] Distilling Multi-Step Reasoning Capabilities into Smaller Language Model
    Yim, Yauwai
    Wang, Zirui
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 530 - 535
  • [9] MusTQ: A Temporal Knowledge Graph Question Answering Dataset for Multi-Step Temporal Reasoning
    Zhang, Tingyi
    Wang, Jiaan
    Li, Zhixu
    Qu, Jianfeng
    Liu, An
    Chen, Zhigang
    Zhi, Hongping
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 11688 - 11699
  • [10] Distilling mathematical reasoning capabilities into Small Language Models
    Zhu, Xunyu
    Li, Jian
    Liu, Yong
    Ma, Can
    Wang, Weiping
    NEURAL NETWORKS, 2024, 179