The Research about Recurrent Model-Agnostic Meta Learning

被引:1
|
作者
Chen, Shaodong [1 ]
Niu, Ziyu [2 ]
机构
[1] Nanyang Inst Technol, Sch Math & Stat, Nanyang, Henan, Peoples R China
[2] Univ Edinburgh, Sch Informat, Artificial Intelligence, Edinburgh, Midlothian, Scotland
关键词
Model-Agnostic Meta Learning; Omniglot dataset; Convolutional Neural Network; Recurrent Neural Network; Long Short-Term Memory; Gated Recurrent Unit; n-way n-shot model;
D O I
10.3103/S1060992X20010075
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Although Deep Neural Networks (DNNs) have performed great success in machine learning domain, they usually show poorly on few-shot learning tasks, where a classifier has to quickly generalize after getting very few samples from each class. A Model-Agnostic Meta Learning (MAML) model, which is able to solve new learning tasks, only using a small number of training data. A MAML model with a Convolutional Neural Network (CNN) architecture is implemented as well, trained on the Omniglot dataset (rather than DNN), as a baseline for image classification tasks. However, our baseline model suffered from a long-period training process and relatively low efficiency. To address these problems, we introduced Recurrent Neural Network (RNN) architecture and its advanced variants into our MAML model, including Long Short-Term Memory (LSTM) architecture and its variants: LSTM-b and Gated Recurrent Unit (GRU). The experiment results, measured by ac- curacies, demonstrate a considerable improvement in image classification performance and training efficiency compared to the baseline models.
引用
收藏
页码:56 / 67
页数:12
相关论文
共 50 条
  • [41] Stochastic Deep Networks with Linear Competing Units for Model-Agnostic Meta-Learning
    Kalais, Konstantinos
    Chatzis, Sotirios
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 10586 - 10597
  • [42] Few-shot RUL estimation based on model-agnostic meta-learning
    Mo, Yu
    Li, Liang
    Huang, Biqing
    Li, Xiu
    JOURNAL OF INTELLIGENT MANUFACTURING, 2023, 34 (05) : 2359 - 2372
  • [43] Few-shot RUL estimation based on model-agnostic meta-learning
    Yu Mo
    Liang Li
    Biqing Huang
    Xiu Li
    Journal of Intelligent Manufacturing, 2023, 34 : 2359 - 2372
  • [44] Domain-Invariant Speaker Vector Projection by Model-Agnostic Meta-Learning
    Kang, Jiawen
    Liu, Ruiqi
    Li, Lantian
    Cai, Yunqi
    Wang, Dong
    Zheng, Thomas Fang
    INTERSPEECH 2020, 2020, : 3825 - 3829
  • [45] Model-Agnostic Metric for Zero-Shot Learning
    Shen, Jiayi
    Wang, Haochen
    Zhang, Anran
    Qiu, Qiang
    Zhen, Xiantong
    Cao, Xianbin
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 775 - 784
  • [46] Meta-LSTM in hydrology: Advancing runoff predictions through model-agnostic meta-learning
    Cai, Kaixuan
    He, Jinxin
    Li, Qingliang
    Wei, Shangguan
    Li, Lu
    Hu, Huiming
    JOURNAL OF HYDROLOGY, 2024, 639
  • [47] Responsible model deployment via model-agnostic uncertainty learning
    Preethi Lahoti
    Krishna Gummadi
    Gerhard Weikum
    Machine Learning, 2023, 112 : 939 - 970
  • [48] Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and Personalized Federated Learning
    Wang, Bokun
    Yuan, Zhuoning
    Ying, Yiming
    Yang, Tianbao
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [49] Learning Model-Agnostic Counterfactual Explanations for Tabular Data
    Pawelczyk, Martin
    Broelemann, Klaus
    Kasneci, Gjergji
    WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, : 3126 - 3132
  • [50] Meta-Attack: Class-agnostic and Model-agnostic Physical Adversarial Attack
    Feng, Weiwei
    Wu, Baoyuan
    Zhang, Tianzhu
    Zhang, Yong
    Zhang, Yongdong
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7767 - 7776