Model Selection of Bayesian Hierarchical Mixture of Experts based on Variational Inference

被引:0
|
作者
Iikubo, Yuji [1 ]
Horii, Shunsuke [2 ]
Matsushima, Toshiyasu [1 ]
机构
[1] Waseda Univ, Dept Appl Math, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
[2] Waseda Univ, Global Educ Ctr, Shinjuku Ku, 1-6-1 Nishiwaseda, Tokyo 1698050, Japan
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We consider the model selection of the hierarchical mixture of experts (HME). The HME is a tree-structured probabilistic model for regression and classification. The HME model has high prediction accuracy and high interpretability, however, the estimation of the parameters tends to overfit due to the complexity of the model. In order to mitigate the overfitting problem, in previous studies, several Bayesian estimation methods for the HME parameters have been proposed. In these studies, the true model that generates data is fixed. In general, however, the true model is unknown. Model selection is one of the most important and difficult problems of regression and classification. For the Bayesian HME, the model is determined by the tree structure, the form of the prior distribution and its parameters, however, only the tree structure is considered as a model parameter in previous studies. In this paper, we consider all of these as model parameters and extend the model selection method. Then, we propose a maximum a posteriori (MAP) estimation method of the Bayesian HME model selection. The approximate posterior probability of each model is calculated by the variational lower bound. We show the effectiveness of the proposed method by numerical experiments and discuss the results applied to actual data sets.
引用
收藏
页码:3474 / 3479
页数:6
相关论文
共 50 条
  • [1] Sparse Bayesian Hierarchical Mixture of Experts and Variational Inference
    Iikubo, Yuji
    Horii, Shunsuke
    Matsushima, Toshiyasu
    [J]. PROCEEDINGS OF 2018 INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY AND ITS APPLICATIONS (ISITA2018), 2018, : 60 - 64
  • [2] Optimal model inference for Bayesian mixture of experts
    Ueda, N
    Ghahramani, Z
    [J]. NEURAL NETWORKS FOR SIGNAL PROCESSING X, VOLS 1 AND 2, PROCEEDINGS, 2000, : 145 - 154
  • [3] Optimal model inference for Bayesian mixture of experts
    Ueda, Naonori
    Ghahramani, Zoubin
    [J]. Neural Networks for Signal Processing - Proceedings of the IEEE Workshop, 2000, 1 : 145 - 154
  • [4] Hierarchical model selection for NGnet based on variational Bayes inference
    Yoshimoto, J
    Ishii, S
    Sato, M
    [J]. ARTIFICIAL NEURAL NETWORKS - ICANN 2002, 2002, 2415 : 661 - 666
  • [5] Bayesian estimation of Dirichlet mixture model with variational inference
    Ma, Zhanyu
    Rana, Pravin Kumar
    Taghia, Jalil
    Flierl, Markus
    Leijon, Arne
    [J]. PATTERN RECOGNITION, 2014, 47 (09) : 3143 - 3157
  • [6] Dirichlet process mixture model based nonparametric Bayesian modeling and variational inference
    Fei, Zhengshun
    Liu, Kangling
    Huang, Bingqiang
    Zheng, Yongping
    Xiang, Xinjian
    [J]. 2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 3048 - 3051
  • [7] SPARSE BAYESIAN HIERARCHICAL MIXTURE OF EXPERTS
    Mossavat, Iman
    Amft, Oliver
    [J]. 2011 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP (SSP), 2011, : 653 - 656
  • [8] Community Embeddings with Bayesian Gaussian Mixture Model and Variational Inference
    Begehr, Anton I. N.
    Panfilov, Peter B.
    [J]. 2022 IEEE 24TH CONFERENCE ON BUSINESS INFORMATICS (CBI 2022), VOL 2, 2022, : 88 - 96
  • [9] Merging experts' opinions: A Bayesian hierarchical model with mixture of prior distributions
    Rufo, M. J.
    Perez, C. J.
    Martin, J.
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2010, 207 (01) : 284 - 289
  • [10] Variational approximations in Bayesian model selection for finite mixture distributions
    McGrory, C. A.
    Titterington, D. M.
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2007, 51 (11) : 5352 - 5367