Learning Symbolic Model-Agnostic Loss Functions via Meta-Learning

被引:1
|
作者
Raymond, Christian [1 ]
Chen, Qi [1 ]
Xue, Bing [1 ]
Zhang, Mengjie [1 ]
机构
[1] Victoria Univ Wellington, Sch Engn & Comp Sci, Evolutionary Computat & Machine Learning Res Grp, Wellington 6140, New Zealand
关键词
Task analysis; Metalearning; Optimization; Training; Market research; Genetic programming; Computer architecture; Loss function learning; meta-learning; evolutionary computation; neuro-symbolic; NEURAL-NETWORKS;
D O I
10.1109/TPAMI.2023.3294394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we develop upon the emerging topic of loss function learning, which aims to learn loss functions that significantly improve the performance of the models trained under them. Specifically, we propose a new meta-learning framework for learning model-agnostic loss functions via a hybrid neuro-symbolic search approach. The framework first uses evolution-based methods to search the space of primitive mathematical operations to find a set of symbolic loss functions. Second, the set of learned loss functions are subsequently parameterized and optimized via an end-to-end gradient-based training procedure. The versatility of the proposed framework is empirically validated on a diverse set of supervised learning tasks. Results show that the meta-learned loss functions discovered by the newly proposed method outperform both the cross-entropy loss and state-of-the-art loss function learning methods on a diverse range of neural network architectures and datasets. We make our code available at *retracted*.
引用
收藏
页码:13699 / 13714
页数:16
相关论文
共 50 条
  • [21] Crop Disease Recognition Based on Improved Model-Agnostic Meta-Learning
    Si, Xiuli
    Hong, Biao
    Hu, Yuanhui
    Chu, Lidong
    [J]. CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 75 (03): : 6101 - 6118
  • [22] Generalization of Model-Agnostic Meta-Learning Algorithms: Recurring and Unseen Tasks
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [23] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [24] A Compressed Model-Agnostic Meta-Learning Model Based on Pruning for Disease Diagnosis
    Hu, Xiangjun
    Ding, Xiuxiu
    Bai, Dongpeng
    Zhang, Qingchen
    [J]. JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2023, 32 (02)
  • [25] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    [J]. Journal of Machine Learning Research, 2022, 23
  • [26] Cross Domain Adaptation of Crowd Counting with Model-Agnostic Meta-Learning
    Hou, Xiaoyu
    Xu, Jihui
    Wu, Jinming
    Xu, Huaiyu
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (24):
  • [27] MODEL-AGNOSTIC META-LEARNING FOR RESILIENCE OPTIMIZATION OF ARTIFICIAL INTELLIGENCE SYSTEM
    Moskalenko, V. V.
    [J]. RADIO ELECTRONICS COMPUTER SCIENCE CONTROL, 2023, (02) : 79 - 90
  • [28] Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and Personalized Federated Learning
    Wang, Bokun
    Yuan, Zhuoning
    Ying, Yiming
    Yang, Tianbao
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [29] On the Convergence Theory of Gradient-Based Model-Agnostic Meta-Learning Algorithms
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 1082 - 1091
  • [30] Specific Emitter Identification With Limited Samples: A Model-Agnostic Meta-Learning Approach
    Yang, Ning
    Zhang, Bangning
    Ding, Guoru
    Wei, Yimin
    Wei, Guofeng
    Wang, Jian
    Guo, Daoxing
    [J]. IEEE COMMUNICATIONS LETTERS, 2022, 26 (02) : 345 - 349