Distilling mathematical reasoning capabilities into Small Language Models

被引:0
|
作者
Zhu, Xunyu [1 ,2 ]
Li, Jian [1 ,2 ]
Liu, Yong [3 ]
Ma, Can [1 ,2 ]
Wang, Weiping [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Large language models; Knowledge Distillation; Mathematical reasoning; Chain-of-Thought; Program-of-Thought;
D O I
10.1016/j.neunet.2024.106594
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work addresses the challenge of democratizing advanced Large Language Models (LLMs) by compressing their mathematical reasoning capabilities into sub-billion parameter Small Language Models (SLMs) without compromising performance. We introduce Equation-of-Thought Distillation (EoTD), a novel technique that encapsulates the reasoning process into equation-based representations to construct an EoTD dataset for finetuning SLMs. Additionally, we propose the Ensemble Thoughts Distillation (ETD) framework to enhance the reasoning performance of SLMs. This involves creating a reasoning dataset with multiple thought processes, including Chain-of-Thought (CoT), Program-of-Thought (PoT), and Equation-of-Thought (EoT), and using it for fine-tuning. Our experimental performance demonstrates that EoTD significantly boosts the reasoning abilities of SLMs, while ETD enables these models to achieve state-of-the-art reasoning performance.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] DISCOURSE GRAMMARS AND STRUCTURE OF MATHEMATICAL REASONING .1. MATHEMATICAL REASONING AND STRATIFICATION OF LANGUAGE
    CORCORAN, J
    JOURNAL OF STRUCTURAL LEARNING, 1971, 3 (01): : 55 - 74
  • [22] OpenToM: A Comprehensive Benchmark for Evaluating Theory-of-Mind Reasoning Capabilities of Large Language Models
    Xu, Hainiu
    Zhao, Runcong
    Zhu, Lixing
    Du, Jinhua
    He, Yulan
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 8593 - 8623
  • [23] The predictive capabilities of mathematical models for the type-token relationship in English language corpora
    Tunnicliffe, Martin
    Hunter, Gordon
    COMPUTER SPEECH AND LANGUAGE, 2021, 70
  • [24] Turning Dust into Gold: Distilling Complex Reasoning Capabilities from LLMs by Leveraging Negative Data
    Li, Yiwei
    Yuan, Peiwen
    Feng, Shaoxiong
    Pan, Boyuan
    Sun, Bin
    Wang, Xinglin
    Wang, Heda
    Li, Kan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18591 - 18599
  • [25] Distilling Script Knowledge from Large Language Models for Constrained Language Planning
    Yuan, Siyu
    Chen, Jiangjie
    Fu, Ziquan
    Ge, Xuyang
    Shah, Soham
    Jankowski, Charles Robert
    Xiao, Yanghua
    Yang, Deqing
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4303 - 4325
  • [26] Reasoning about mathematical models - Preface
    Greenberg, HJ
    ANNALS OF MATHEMATICS AND ARTIFICIAL INTELLIGENCE, 1996, 17 (1-2) : U9 - U9
  • [27] TempLM: Distilling Language Models into Template-Based Generators
    Zhang, Tianyi
    Lee, Mina
    Li, Lisa
    Shen, Ende
    Hashimoto, Tatsunori B.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1970 - 1994
  • [28] LLM performance on mathematical reasoning in Catalan language
    Rhomrasi, Lamyae
    Ahsini, Yusef
    Igualde-Saez, Arnau
    Vinuesa, Ricardo
    Hoyas, Sergio
    Garcia-Sabater, Jose Pedro
    Fullana-i-Alfonso, Marius J.
    Conejero, J. Alberto
    RESULTS IN ENGINEERING, 2025, 25
  • [29] tinyCLAP: Distilling Contrastive Language-Audio Pretrained models
    Paissan, Francesco
    Farella, Elisabetta
    INTERSPEECH 2024, 2024, : 1685 - 1689
  • [30] From Complex to Simple: Unraveling the Cognitive Tree for Reasoning with Small Language Models
    Yan, Junbing
    Wang, Chengyu
    Zhang, Taolin
    He, Xiaofeng
    Huang, Jun
    Zhang, Wei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12413 - 12425