Distilling mathematical reasoning capabilities into Small Language Models

被引:0
|
作者
Zhu, Xunyu [1 ,2 ]
Li, Jian [1 ,2 ]
Liu, Yong [3 ]
Ma, Can [1 ,2 ]
Wang, Weiping [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Large language models; Knowledge Distillation; Mathematical reasoning; Chain-of-Thought; Program-of-Thought;
D O I
10.1016/j.neunet.2024.106594
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work addresses the challenge of democratizing advanced Large Language Models (LLMs) by compressing their mathematical reasoning capabilities into sub-billion parameter Small Language Models (SLMs) without compromising performance. We introduce Equation-of-Thought Distillation (EoTD), a novel technique that encapsulates the reasoning process into equation-based representations to construct an EoTD dataset for finetuning SLMs. Additionally, we propose the Ensemble Thoughts Distillation (ETD) framework to enhance the reasoning performance of SLMs. This involves creating a reasoning dataset with multiple thought processes, including Chain-of-Thought (CoT), Program-of-Thought (PoT), and Equation-of-Thought (EoT), and using it for fine-tuning. Our experimental performance demonstrates that EoTD significantly boosts the reasoning abilities of SLMs, while ETD enables these models to achieve state-of-the-art reasoning performance.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Distilling Reasoning Capabilities into Smaller Language Models
    Shridhar, Kumar
    Stolfo, Alessandro
    Sachan, Mrinmaya
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 7059 - 7073
  • [2] Distilling Multi-Step Reasoning Capabilities into Smaller Language Model
    Yim, Yauwai
    Wang, Zirui
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 530 - 535
  • [3] Disentangling Reasoning Capabilities from Language Models with Compositional Reasoning Transformers
    Zhong, Wanjun
    Mae, Tingting
    Wang, Jiahai
    Yin, Jian
    Zhao, Tiejun
    Lin, Chin-Yew
    Duan, Nan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 7587 - 7600
  • [4] Large Language Models for Mathematical Reasoning: Progresses and Challenges
    Ahn, Janice
    Verma, Rishu
    Lou, Renze
    Zhang, Rui
    Yin, Wenpeng
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: STUDENT RESEARCH WORKSHOP, 2024, : 225 - 237
  • [5] Visual Program Distillation: Distilling Tools and Programmatic Reasoning into Vision-Language Models
    Hu, Yushi
    Stretcu, Otilia
    Lu, Chun-Ta
    Viswanathan, Krishnamurthy
    Hata, Kenji
    Luo, Enming
    Krishna, Ranjay
    Fuxman, Ariel
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 9590 - 9601
  • [6] Mathematical Reasoning via Multi-step Self Questioning and Answering for Small Language Models
    Chen, Kaiyuan
    Wang, Jin
    Zhang, Xuejie
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT IV, NLPCC 2024, 2025, 15362 : 81 - 93
  • [7] An Evaluation of Reasoning Capabilities of Large Language Models in Financial Sentiment Analysis
    Du, Kelvin
    Xing, Frank
    Mao, Rui
    Cambria, Erik
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 189 - 194
  • [8] CommonsenseVIS: Visualizing and Understanding Commonsense Reasoning Capabilities of Natural Language Models
    Wang, Xingbo
    Huang, Renfei
    Jin, Zhihua
    Fang, Tianqing
    Qu, Huamin
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (01) : 273 - 283
  • [9] A Causal Framework to Quantify the Robustness of Mathematical Reasoning with Language Models
    Stolfo, Alessandro
    Jin, Zhijing
    Shridhar, Kumar
    Scholkopf, Bernhard
    Sachan, Mrinmaya
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 545 - 561
  • [10] Exploring Reversal Mathematical Reasoning Ability for Large Language Models
    Guo, Pei
    You, Wangjie
    Li, Juntao
    Yan, Bowen
    Zhang, Min
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 13671 - 13685