Probabilistic Model-Agnostic Meta-Learning

被引:0
|
作者
Finn, Chelsea [1 ]
Xu, Kelvin [1 ]
Levine, Sergey [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning for few-shot learning entails acquiring a prior over previous tasks and experiences, such that new tasks be learned from small amounts of data. However, a critical challenge in few-shot learning is task ambiguity: even when a powerful prior can be meta-learned from a large number of prior tasks, a small dataset for a new task can simply be too ambiguous to acquire a single model (e.g., a classifier) for that task that is accurate. In this paper, we propose a probabilistic meta-learning algorithm that can sample models for a new task from a model distribution. Our approach extends model-agnostic meta-learning, which adapts to new tasks via gradient descent, to incorporate a parameter distribution that is trained via a variational lower bound. At meta-test time, our algorithm adapts via a simple procedure that injects noise into gradient descent, and at meta-training time, the model is trained such that this stochastic adaptation procedure produces samples from the approximate model posterior. Our experimental results show that our method can sample plausible classifiers and regressors in ambiguous few-shot learning problems. We also show how reasoning about ambiguity can also be used for downstream active learning problems.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Visual analysis of meteorological satellite data via model-agnostic meta-learning
    Shiyu Cheng
    Hanwei Shen
    Guihua Shan
    Beifang Niu
    Weihua Bai
    [J]. Journal of Visualization, 2021, 24 : 301 - 315
  • [22] Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation
    Vuorio, Risto
    Sun, Shao-Hua
    Hu, Hexiang
    Lim, Joseph J.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [23] On the Convergence Theory of Gradient-Based Model-Agnostic Meta-Learning Algorithms
    Fallah, Alireza
    Mokhtari, Aryan
    Ozdaglar, Asuman
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 1082 - 1091
  • [24] Specific Emitter Identification With Limited Samples: A Model-Agnostic Meta-Learning Approach
    Yang, Ning
    Zhang, Bangning
    Ding, Guoru
    Wei, Yimin
    Wei, Guofeng
    Wang, Jian
    Guo, Daoxing
    [J]. IEEE COMMUNICATIONS LETTERS, 2022, 26 (02) : 345 - 349
  • [25] Visual analysis of meteorological satellite data via model-agnostic meta-learning
    Cheng, Shiyu
    Shen, Hanwei
    Shan, Guihua
    Niu, Beifang
    Bai, Weihua
    [J]. JOURNAL OF VISUALIZATION, 2021, 24 (02) : 301 - 315
  • [26] Few-shot RUL estimation based on model-agnostic meta-learning
    Mo, Yu
    Li, Liang
    Huang, Biqing
    Li, Xiu
    [J]. JOURNAL OF INTELLIGENT MANUFACTURING, 2023, 34 (05) : 2359 - 2372
  • [27] Domain-Invariant Speaker Vector Projection by Model-Agnostic Meta-Learning
    Kang, Jiawen
    Liu, Ruiqi
    Li, Lantian
    Cai, Yunqi
    Wang, Dong
    Zheng, Thomas Fang
    [J]. INTERSPEECH 2020, 2020, : 3825 - 3829
  • [28] Few-shot RUL estimation based on model-agnostic meta-learning
    Yu Mo
    Liang Li
    Biqing Huang
    Xiu Li
    [J]. Journal of Intelligent Manufacturing, 2023, 34 : 2359 - 2372
  • [29] Stochastic Deep Networks with Linear Competing Units for Model-Agnostic Meta-Learning
    Kalais, Konstantinos
    Chatzis, Sotirios
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 10586 - 10597
  • [30] Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and Personalized Federated Learning
    Wang, Bokun
    Yuan, Zhuoning
    Ying, Yiming
    Yang, Tianbao
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24