A Recursive tree-structured neural network with goal forgetting and information aggregation for solving math word problems

被引:2
|
作者
Xiao, Jing [1 ]
Huang, Linjia [1 ]
Song, Yu [2 ]
Tang, Na [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou 510631, Peoples R China
[2] South China Normal Univ, Sch Educ, Guangzhou 510631, Peoples R China
基金
中国国家自然科学基金;
关键词
Math word problems; Deep learning; Tree-Structured decoder;
D O I
10.1016/j.ipm.2023.103324
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Most existing state-of-the-art neural network models for math word problems use the Goal -driven Tree-Structured decoder (GTS) to generate expression trees. However, we found that GTS does not provide good predictions for longer expressions, mainly because it does not capture the relationships among the goal vectors of each node in the expression tree and ignores the position order of the nodes before and after the operator. In this paper, we propose a novel Recursive tree-structured neural network with Goal Forgetting and information aggregation (RGFNet) to address these limits. The goal forgetting and information aggregation module is based on ordinary differential equations (ODEs) and we use it to build a sub-goal information feedback neural network (SGIFNet). Unlike GTS, which uses two-layer gated-feedforward networks to generate goal vectors, we introduce a novel sub-goal generation module. The sub-goal generation module could capture the relationship among the related nodes (e.g. parent nodes, sibling nodes) using attention mechanism. Experimental results on two large public datasets i.e. Math23K and Ape-clean show that our tree-structured model outperforms the state-of-the-art models and obtains answer accuracy over 86%. Furthermore, the performance on long-expression problems is promising.1
引用
收藏
页数:16
相关论文
共 34 条
  • [1] A Goal-Driven Tree-Structured Neural Model for Math Word Problems
    Xie, Zhipeng
    Sun, Shichao
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5299 - 5305
  • [2] Semantically-Aligned Universal Tree-Structured Solver for Math Word Problems
    Qin, Jinghui
    Lin, Lihui
    Liang, Xiaodan
    Zhang, Rumin
    Lin, Liang
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3780 - 3789
  • [3] Tree-structured Decoding for Solving MathWord Problems
    Liu, Qianying
    Guan, Wenyu
    Li, Sujian
    Kawahara, Daisuke
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 2370 - 2379
  • [4] Goal selection and feedback for solving math word problems
    He, Daijun
    Xiao, Jing
    [J]. APPLIED INTELLIGENCE, 2023, 53 (12) : 14744 - 14758
  • [5] Goal selection and feedback for solving math word problems
    Daijun He
    Jing Xiao
    [J]. Applied Intelligence, 2023, 53 : 14744 - 14758
  • [6] Tree-structured multilayer neural network for classification
    Shiueng-Bien Yang
    [J]. Neural Computing and Applications, 2020, 32 : 5859 - 5873
  • [7] Rumor Detection on Twitter with Tree-structured Recursive Neural Networks
    Ma, Jing
    Gao, Wei
    Wong, Kam-Fai
    [J]. PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, 2018, : 1980 - 1989
  • [8] Tree-Structured Neural Network for Hyperspectral Pansharpening
    He, Lin
    Ye, Hanghui
    Xi, Dahan
    Li, Jun
    Plaza, Antonio
    Zhang, Mei
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 2516 - 2530
  • [9] Tree-structured multilayer neural network for classification
    Yang, Shiueng-Bien
    [J]. NEURAL COMPUTING & APPLICATIONS, 2020, 32 (10): : 5859 - 5873
  • [10] Graph-to-Tree Learning for Solving Math Word Problems
    Zhang, Jipeng
    Wang, Lei
    Lee, Roy Ka-Wei
    Bin, Yi
    Wang, Yan
    Shao, Jie
    Lim, Ee-Peng
    [J]. 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3928 - 3937