Template-Based Contrastive Distillation Pretraining for Math Word Problem Solving

被引:0
|
作者
Qin, Jinghui [1 ]
Yang, Zhicheng [2 ,3 ]
Chen, Jiaqi [1 ]
Liang, Xiaodan [2 ]
Lin, Liang [1 ]
机构
[1] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou 510275, Peoples R China
[2] Sun Yat Sen Univ, Sch Intelligent Syst Engn, Shenzhen Campus, Shenzhen, Peoples R China
[3] Dark Matter Inc, Guangzhou 511457, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Mathematical models; Task analysis; Semantics; Problem-solving; Linguistics; Representation learning; Predictive models; Contrastive learning; math word problem (MWP) solving automatically; model pretraining; natural language understanding;
D O I
10.1109/TNNLS.2023.3265173
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since math word problem (MWP) solving aims to transform natural language problem description into executable solution equations, an MWP solver needs to not only comprehend the real-world narrative described in the problem text but also identify the relationships among the quantifiers and variables implied in the problem and maps them into a reasonable solution equation logic. Recently, although deep learning models have made great progress in MWPs, they ignore the grounding equation logic implied by the problem text. Besides, as we all know, pretrained language models (PLM) have a wealth of knowledge and high-quality semantic representations, which may help solve MWPs, but they have not been explored in the MWP-solving task. To harvest the equation logic and real-world knowledge, we propose a template-based contrastive distillation pretraining (TCDP) approach based on a PLM-based encoder to incorporate mathematical logic knowledge by multiview contrastive learning while retaining rich real-world knowledge and high-quality semantic representation via knowledge distillation. We named the pretrained PLM-based encoder by our approach as MathEncoder. Specifically, the mathematical logic is first summarized by clustering the symbolic solution templates among MWPs and then injected into the deployed PLM-based encoder by conducting supervised contrastive learning based on the symbolic solution templates, which can represent the underlying solving logic in the problems. Meanwhile, the rich knowledge and high-quality semantic representation are retained by distilling them from a well-trained PLM-based teacher encoder into our MathEncoder. To validate the effectiveness of our pretrained MathEncoder, we construct a new solver named MathSolver by replacing the GRU-based encoder with our pretrained MathEncoder in GTS, which is a state-of-the-art MWP solver. The experimental results demonstrate that our method can carry a solver's understanding ability of MWPs to a new stage by outperforming existing state-of-the-art methods on two widely adopted benchmarks Math23K and CM17K. Code will be available at https://github.com/QinJinghui/tcdp.
引用
收藏
页码:12823 / 12835
页数:13
相关论文
共 50 条
  • [21] Teacher-Student Networks with Multiple Decoders for Solving Math Word Problem
    Zhang, Jipeng
    Lee, Roy Ka-Wei
    Lim, Ee-Peng
    Qin, Wei
    Wang, Lei
    Shao, Jie
    Sun, Qianru
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 4011 - 4017
  • [22] The Effects of Dynamic Strategic Math on English Language Learners' Word Problem Solving
    Orosco, Michael J.
    Swanson, H. Lee
    O'Connor, Rollanda
    Lussier, Cathy
    JOURNAL OF SPECIAL EDUCATION, 2013, 47 (02): : 96 - 107
  • [23] Recycling Numeracy Data Augmentation with Symbolic Verification for Math Word Problem Solving
    Jen, Tien-Yi
    Huang, Hen-Hsen
    Chen, Hsin-Hsi
    2021 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY (WI-IAT 2021), 2021, : 653 - 657
  • [24] Learning to Reason Deductively: Math Word Problem Solving as Complex Relation Extraction
    Jie, Zhanming
    Li, Jierui
    Lu, Wei
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5944 - 5955
  • [25] Word-Problem-Solving Strategy for Minority Students at Risk for Math Difficulties
    Kong, Jennifer E.
    Orosco, Michael J.
    LEARNING DISABILITY QUARTERLY, 2016, 39 (03) : 171 - 181
  • [26] Learning Dynamic Contextualised Word Embeddings via Template-based Temporal Adaptation
    Tang, Xiaohang
    Zhou, Yi
    Bollegala, Danushka
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 9352 - 9369
  • [27] Incorporating Template-Based Contrastive Learning into Cognitively Inspired, Low-Resource Relation Extraction
    Zheng, Yandan
    Tuan, Luu Anh
    COGNITIVE COMPUTATION, 2024, 16 (06) : 3228 - 3240
  • [28] Arithmetic Word Problem-Solving and Math Anxiety: The Role of Perceived Difficulty and Gender
    Doz, Eleonora
    Cuder, Alessandro
    Pellizzoni, Sandra
    Carretti, Barbara
    Passolunghi, Maria Chiara
    JOURNAL OF COGNITION AND DEVELOPMENT, 2023, 24 (04) : 598 - 616
  • [29] Word Problem Solving in Contemporary Math Education: A Plea for Reading Comprehension Skills Training
    Boonen, Anton J. H.
    de Koning, Bjorn B.
    Jolles, Jelle
    van der Schoot, Menno
    FRONTIERS IN PSYCHOLOGY, 2016, 7
  • [30] A Knowledge-Aware Sequence-to-Tree Network for Math Word Problem Solving
    Wu, Qinzhuo
    Zhang, Qi
    Fu, Jinlan
    Huang, Xuanjing
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7137 - 7146