Learning Symbolic Model-Agnostic Loss Functions via Meta-Learning

被引:1
|
作者
Raymond, Christian [1 ]
Chen, Qi [1 ]
Xue, Bing [1 ]
Zhang, Mengjie [1 ]
机构
[1] Victoria Univ Wellington, Sch Engn & Comp Sci, Evolutionary Computat & Machine Learning Res Grp, Wellington 6140, New Zealand
关键词
Task analysis; Metalearning; Optimization; Training; Market research; Genetic programming; Computer architecture; Loss function learning; meta-learning; evolutionary computation; neuro-symbolic; NEURAL-NETWORKS;
D O I
10.1109/TPAMI.2023.3294394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we develop upon the emerging topic of loss function learning, which aims to learn loss functions that significantly improve the performance of the models trained under them. Specifically, we propose a new meta-learning framework for learning model-agnostic loss functions via a hybrid neuro-symbolic search approach. The framework first uses evolution-based methods to search the space of primitive mathematical operations to find a set of symbolic loss functions. Second, the set of learned loss functions are subsequently parameterized and optimized via an end-to-end gradient-based training procedure. The versatility of the proposed framework is empirically validated on a diverse set of supervised learning tasks. Results show that the meta-learned loss functions discovered by the newly proposed method outperform both the cross-entropy loss and state-of-the-art loss function learning methods on a diverse range of neural network architectures and datasets. We make our code available at *retracted*.
引用
收藏
页码:13699 / 13714
页数:16
相关论文
共 50 条
  • [1] Meta weight learning via model-agnostic meta-learning
    Xu, Zhixiong
    Chen, Xiliang
    Tang, Wei
    Lai, Jun
    Cao, Lei
    [J]. NEUROCOMPUTING, 2021, 432 : 124 - 132
  • [2] Probabilistic Model-Agnostic Meta-Learning
    Finn, Chelsea
    Xu, Kelvin
    Levine, Sergey
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] Bayesian Model-Agnostic Meta-Learning
    Yoon, Jaesik
    Kim, Taesup
    Dia, Ousmane
    Kim, Sungwoong
    Bengio, Yoshua
    Ahn, Sungjin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [4] Knowledge Distillation for Model-Agnostic Meta-Learning
    Zhang, Min
    Wang, Donglin
    Gai, Sibo
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1355 - 1362
  • [5] Combining Model-Agnostic Meta-Learning and Transfer Learning for Regression
    Satrya, Wahyu Fadli
    Yun, Ji-Hoon
    [J]. SENSORS, 2023, 23 (02)
  • [6] Task-Robust Model-Agnostic Meta-Learning
    Collins, Liam
    Mokhtari, Aryan
    Shakkottai, Sanjay
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [7] Visual analysis of meteorological satellite data via model-agnostic meta-learning
    Shiyu Cheng
    Hanwei Shen
    Guihua Shan
    Beifang Niu
    Weihua Bai
    [J]. Journal of Visualization, 2021, 24 : 301 - 315
  • [8] Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation
    Vuorio, Risto
    Sun, Shao-Hua
    Hu, Hexiang
    Lim, Joseph J.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Visual analysis of meteorological satellite data via model-agnostic meta-learning
    Cheng, Shiyu
    Shen, Hanwei
    Shan, Guihua
    Niu, Beifang
    Bai, Weihua
    [J]. JOURNAL OF VISUALIZATION, 2021, 24 (02) : 301 - 315
  • [10] Specific Emitter Identification via Sparse Bayesian Learning Versus Model-Agnostic Meta-Learning
    He, Boxiang
    Wang, Fanggang
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2023, 18 : 3677 - 3691