Hierarchical Gaussian Mixture based Task Generative Model for Robust Meta-Learning

被引:0
|
作者
Zhang, Yizhou [1 ]
Ni, Jingchao [2 ]
Cheng, Wei [3 ]
Chen, Zhengzhang [3 ]
Tong, Liang [4 ]
Chen, Haifeng [3 ]
Liu, Yan [1 ]
机构
[1] Univ Southern Calif, Los Angeles, CA 90007 USA
[2] AWS AI Labs, Seattle, WA USA
[3] NEC Labs Amer, Irving, TX USA
[4] Stellar Cyber Inc, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning enables quick adaptation of machine learning models to new tasks with limited data. While tasks could come from varying distributions in reality, most of the existing meta-learning methods consider both training and testing tasks as from the same uni-component distribution, overlooking two critical needs of a practical solution: (1) the various sources of tasks may compose a multi-component mixture distribution, and (2) novel tasks may come from a distribution that is unseen during meta-training. In this paper, we demonstrate these two challenges can be solved jointly by modeling the density of task instances. We develop a metatraining framework underlain by a novel Hierarchical Gaussian Mixture based Task Generative Model (HTGM). HTGM extends the widely used empirical process of sampling tasks to a theoretical model, which learns task embeddings, fits the mixture distribution of tasks, and enables density-based scoring of novel tasks. The framework is agnostic to the encoder and scales well with large backbone networks. The model parameters are learned end-to-end by maximum likelihood estimation via an Expectation-Maximization (EM) algorithm. Extensive experiments on benchmark datasets indicate the effectiveness of our method for both sample classification and novel task detection.
引用
收藏
页数:24
相关论文
共 50 条
  • [1] Task-Robust Model-Agnostic Meta-Learning
    Collins, Liam
    Mokhtari, Aryan
    Shakkottai, Sanjay
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [2] GAME: GAussian Mixture Error-based meta-learning architecture
    Jinhe Dong
    Jun Shi
    Yue Gao
    Shihui Ying
    Neural Computing and Applications, 2023, 35 : 20445 - 20461
  • [3] GAME: GAussian Mixture Error-based meta-learning architecture
    Dong, Jinhe
    Shi, Jun
    Gao, Yue
    Ying, Shihui
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (28): : 20445 - 20461
  • [4] Task Aligned Generative Meta-learning for Zero-shot Learning
    Liu, Zhe
    Li, Yun
    Yao, Lina
    Wang, Xianzhi
    Long, Guodong
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8723 - 8731
  • [5] Task-based Focal Loss for Adversarially Robust Meta-Learning
    Hou, Yufan
    Zou, Lixin
    Liu, Weidong
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2824 - 2829
  • [6] Generative Meta-Learning Robust Quality-Diversity Portfolio
    Yuksel, Kamer Ali
    PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION, 2023, : 787 - 790
  • [7] Meta-learning representations for clustering with infinite Gaussian mixture models
    Iwata, Tomoharu
    NEUROCOMPUTING, 2023, 549
  • [8] The meta-learning method for the ensemble model based on situational meta-task
    Zhang, Zhengchao
    Zhou, Lianke
    Wu, Yuyang
    Wang, Nianbin
    FRONTIERS IN NEUROROBOTICS, 2024, 18
  • [9] ROBUST MAML: PRIORITIZATION TASK BUFFER WITH ADAPTIVE LEARNING PROCESS FOR MODEL-AGNOSTIC META-LEARNING
    Thanh Nguyen
    Tung Luu
    Trung Pham
    Rakhimkul, Sanzhar
    Yoo, Chang D.
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3460 - 3464
  • [10] Scalable Meta-Learning with Gaussian Processes
    Tighineanu, Petru
    Grossberger, Lukas
    Baireuther, Paul
    Skubch, Kathrin
    Falkner, Stefan
    Vinogradska, Julia
    Berkenkamp, Felix
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238