The Research about Recurrent Model-Agnostic Meta Learning

被引:1
|
作者
Chen, Shaodong [1 ]
Niu, Ziyu [2 ]
机构
[1] Nanyang Inst Technol, Sch Math & Stat, Nanyang, Henan, Peoples R China
[2] Univ Edinburgh, Sch Informat, Artificial Intelligence, Edinburgh, Midlothian, Scotland
关键词
Model-Agnostic Meta Learning; Omniglot dataset; Convolutional Neural Network; Recurrent Neural Network; Long Short-Term Memory; Gated Recurrent Unit; n-way n-shot model;
D O I
10.3103/S1060992X20010075
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Although Deep Neural Networks (DNNs) have performed great success in machine learning domain, they usually show poorly on few-shot learning tasks, where a classifier has to quickly generalize after getting very few samples from each class. A Model-Agnostic Meta Learning (MAML) model, which is able to solve new learning tasks, only using a small number of training data. A MAML model with a Convolutional Neural Network (CNN) architecture is implemented as well, trained on the Omniglot dataset (rather than DNN), as a baseline for image classification tasks. However, our baseline model suffered from a long-period training process and relatively low efficiency. To address these problems, we introduced Recurrent Neural Network (RNN) architecture and its advanced variants into our MAML model, including Long Short-Term Memory (LSTM) architecture and its variants: LSTM-b and Gated Recurrent Unit (GRU). The experiment results, measured by ac- curacies, demonstrate a considerable improvement in image classification performance and training efficiency compared to the baseline models.
引用
收藏
页码:56 / 67
页数:12
相关论文
共 50 条
  • [21] Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
    Finn, Chelsea
    Abbeel, Pieter
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [22] Personalized Federated Learning with Theoretical Guarantees: A Model-Agnostic Meta-Learning Approach
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [23] MAME : Model-Agnostic Meta-Exploration
    Gurumurthy, Swaminathan
    Kumar, Sumit
    Sycara, Katia
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [24] A Compressed Model-Agnostic Meta-Learning Model Based on Pruning for Disease Diagnosis
    Hu, Xiangjun
    Ding, Xiuxiu
    Bai, Dongpeng
    Zhang, Qingchen
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2023, 32 (02)
  • [25] Dynamic Model-Agnostic Meta-Learning for Incremental Few-Shot Learning
    Domoguen, Jansen Keith L.
    Naval, Prospero C., Jr.
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4927 - 4933
  • [26] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [27] Crop Disease Recognition Based on Improved Model-Agnostic Meta-Learning
    Si, Xiuli
    Hong, Biao
    Hu, Yuanhui
    Chu, Lidong
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 75 (03): : 6101 - 6118
  • [28] Generalization of Model-Agnostic Meta-Learning Algorithms: Recurring and Unseen Tasks
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [29] Sharp-MAML: Sharpness-Aware Model-Agnostic Meta Learning
    Abbas, Momin
    Xiao, Quan
    Chen, Lisha
    Chen, Pin-Yu
    Chen, Tianyi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 10 - 32
  • [30] MODEL-AGNOSTIC META-LEARNING FOR RESILIENCE OPTIMIZATION OF ARTIFICIAL INTELLIGENCE SYSTEM
    Moskalenko, V. V.
    RADIO ELECTRONICS COMPUTER SCIENCE CONTROL, 2023, (02) : 79 - 90