Hierarchical Gaussian Mixture based Task Generative Model for Robust Meta-Learning

被引:0
|
作者
Zhang, Yizhou [1 ]
Ni, Jingchao [2 ]
Cheng, Wei [3 ]
Chen, Zhengzhang [3 ]
Tong, Liang [4 ]
Chen, Haifeng [3 ]
Liu, Yan [1 ]
机构
[1] Univ Southern Calif, Los Angeles, CA 90007 USA
[2] AWS AI Labs, Seattle, WA USA
[3] NEC Labs Amer, Irving, TX USA
[4] Stellar Cyber Inc, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning enables quick adaptation of machine learning models to new tasks with limited data. While tasks could come from varying distributions in reality, most of the existing meta-learning methods consider both training and testing tasks as from the same uni-component distribution, overlooking two critical needs of a practical solution: (1) the various sources of tasks may compose a multi-component mixture distribution, and (2) novel tasks may come from a distribution that is unseen during meta-training. In this paper, we demonstrate these two challenges can be solved jointly by modeling the density of task instances. We develop a metatraining framework underlain by a novel Hierarchical Gaussian Mixture based Task Generative Model (HTGM). HTGM extends the widely used empirical process of sampling tasks to a theoretical model, which learns task embeddings, fits the mixture distribution of tasks, and enables density-based scoring of novel tasks. The framework is agnostic to the encoder and scales well with large backbone networks. The model parameters are learned end-to-end by maximum likelihood estimation via an Expectation-Maximization (EM) algorithm. Extensive experiments on benchmark datasets indicate the effectiveness of our method for both sample classification and novel task detection.
引用
收藏
页数:24
相关论文
共 50 条
  • [41] Multimodal meta-learning through meta-learned task representations
    Vettoruzzo, Anna
    Bouguelia, Mohamed-Rafik
    Rognvaldsson, Thorsteinn
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (15): : 8519 - 8529
  • [42] Meta-learning: searching in the model space
    Duch, W
    Grudzinski, K
    8TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING, VOLS 1-3, PROCEEDING, 2001, : 235 - 240
  • [43] DeepGUM: Learning Deep Robust Regression with a Gaussian-Uniform Mixture Model
    Lathuiliere, Stephane
    Mesejo, Pablo
    Alameda-Pineda, Xavier
    Horaud, Radu
    COMPUTER VISION - ECCV 2018, PT V, 2018, 11209 : 205 - 221
  • [44] Multimodal Model-Agnostic Meta-Learning via Task-Aware Modulation
    Vuorio, Risto
    Sun, Shao-Hua
    Hu, Hexiang
    Lim, Joseph J.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [45] Gaussian mixture learning via adaptive hierarchical clustering
    Li, Jichuan
    Nehorai, Arye
    SIGNAL PROCESSING, 2018, 150 : 116 - 121
  • [46] A META-ANALYSIS FRAMEWORK BASED ON HIERARCHICAL MIXTURE MODEL
    Dimitri, A.
    Talamo, M.
    ADVANCES AND APPLICATIONS IN STATISTICS, 2007, 7 (03) : 341 - 356
  • [47] ST-MAML : A Stochastic-Task based Method for Task-Heterogeneous Meta-Learning
    Wang, Zhe
    Grigsby, Jake
    Sekhon, Arshdeep
    Qi, Yanjun
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 180, 2022, 180 : 2066 - 2074
  • [48] MAML2: meta reinforcement learning via meta-learning for task categories
    Fu, Qiming
    Wang, Zhechao
    Fang, Nengwei
    Xing, Bin
    Zhang, Xiao
    Chen, Jianping
    FRONTIERS OF COMPUTER SCIENCE, 2023, 17 (04)
  • [49] Symplectic Neural Gaussian Processes for Meta-learning Hamiltonian Dynamics
    Iwata, Tomoharu
    Tanaka, Yusuke
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 4210 - 4218
  • [50] Robust image reconstruction enhancement based on Gaussian mixture model estimation
    Zhao, Fan
    Zhao, Jian
    Han, Xizhen
    Wang, He
    Liu, Bochao
    JOURNAL OF ELECTRONIC IMAGING, 2016, 25 (02)