Evolutionary Neural Architecture Search for Transformer in Knowledge Tracing

被引:0
|
作者
Yang, Shangshang [1 ,3 ]
Yu, Xiaoshan [1 ,3 ]
Tian, Ye [1 ,3 ]
Yan, Xueming [2 ]
Ma, Haiping [1 ,3 ,4 ]
Zhang, Xingyi [1 ,3 ]
机构
[1] Anhui Univ, Hefei, Peoples R China
[2] Guangdong Univ Foreign Studies, Guangzhou, Peoples R China
[3] Minist Educ, Key Lab Intelligent Comp & Signal Proc, Hefei, Peoples R China
[4] Dept Informat Mat & Intelligent Sensing Lab Anhui, Hefei, Peoples R China
基金
中国博士后科学基金; 国家重点研发计划; 中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer has achieved excellent performance in the knowledge tracing (KT) task, but they are criticized for the manually selected input features for fusion and the defect of single global context modelling to directly capture students' forgetting behavior in KT, when the related records are distant from the current record in terms of time. To address the issues, this paper first considers adding convolution operations to the Transformer to enhance its local context modelling ability used for students' forgetting behavior, then proposes an evolutionary neural architecture search approach to automate the input feature selection and automatically determine where to apply which operation for achieving the balancing of the local/global context modelling. In the search space design, the original global path containing the attention module in Transformer is replaced with the sum of a global path and a local path that could contain different convolutions, and the selection of input features is also considered. To search the best architecture, we employ an effective evolutionary algorithm to explore the search space and also suggest a search space reduction strategy to accelerate the convergence of the algorithm. Experimental results on the two largest and most challenging education datasets demonstrate the effectiveness of the architecture found by the proposed approach.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Knowledge transfer evolutionary search for lightweight neural architecture with dynamic inference
    Qian, Xiaoxue
    Liu, Fang
    Jiao, Licheng
    Zhang, Xiangrong
    Huang, Xinyan
    Li, Shuo
    Chen, Puhua
    Liu, Xu
    [J]. PATTERN RECOGNITION, 2023, 143
  • [2] Knowledge reconstruction assisted evolutionary algorithm for neural network architecture search
    An, Yang
    Zhang, Changsheng
    Zheng, Xuanyu
    [J]. KNOWLEDGE-BASED SYSTEMS, 2023, 264
  • [3] An approach for combining multimodal fusion and neural architecture search applied to knowledge tracing
    Xinyi Ding
    Tao Han
    Yili Fang
    Eric Larson
    [J]. Applied Intelligence, 2023, 53 : 11092 - 11103
  • [4] An approach for combining multimodal fusion and neural architecture search applied to knowledge tracing
    Ding, Xinyi
    Han, Tao
    Fang, Yili
    Larson, Eric
    [J]. APPLIED INTELLIGENCE, 2023, 53 (09) : 11092 - 11103
  • [5] Evolutionary neural architecture search combining multi-branch ConvNet and improved transformer
    Yang Xu
    Yongjie Ma
    [J]. Scientific Reports, 13
  • [6] Evolutionary neural architecture search combining multi-branch ConvNet and improved transformer
    Xu, Yang
    Ma, Yongjie
    [J]. SCIENTIFIC REPORTS, 2023, 13 (01)
  • [7] Evolutionary approximation and neural architecture search
    Pinos, Michal
    Mrazek, Vojtech
    Sekanina, Lukas
    [J]. GENETIC PROGRAMMING AND EVOLVABLE MACHINES, 2022, 23 (03) : 351 - 374
  • [8] A Survey on Evolutionary Neural Architecture Search
    Liu, Yuqiao
    Sun, Yanan
    Xue, Bing
    Zhang, Mengjie
    Yen, Gary G.
    Tan, Kay Chen
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (02) : 550 - 570
  • [9] Evolutionary Neural Architecture Search and Applications
    Sun, Yanan
    Zhang, Mengjie
    Yen, Gary G.
    [J]. IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2021, 16 (03) : 8 - 9
  • [10] Evolutionary Recurrent Neural Architecture Search
    Tian, Shuo
    Hu, Kai
    Guo, Shasha
    Li, Shiming
    Wang, Lei
    Xu, Weixia
    [J]. IEEE EMBEDDED SYSTEMS LETTERS, 2021, 13 (03) : 110 - 113