A similarity-based Bayesian mixture-of-experts model

被引:0
|
作者
Tianfang Zhang
Rasmus Bokrantz
Jimmy Olsson
机构
[1] KTH Royal Institute of Technology,Department of Mathematics
[2] RaySearch Laboratories,undefined
[3] Silo AI,undefined
来源
Statistics and Computing | 2023年 / 33卷
关键词
Mixture-of-experts; Nonparametric Bayesian regression; -nearest neighbors; Pseudolikelihood; Variational inference; Reparameterization trick;
D O I
暂无
中图分类号
学科分类号
摘要
We present a new nonparametric mixture-of-experts model for multivariate regression problems, inspired by the probabilistic k-nearest neighbors algorithm. Using a conditionally specified model, predictions for out-of-sample inputs are based on similarities to each observed data point, yielding predictive distributions represented by Gaussian mixtures. Posterior inference is performed on the parameters of the mixture components as well as the distance metric using a mean-field variational Bayes algorithm accompanied with a stochastic gradient-based optimization procedure. The proposed method is especially advantageous in settings where inputs are of relatively high dimension in comparison to the data size, where input–output relationships are complex, and where predictive distributions may be skewed or multimodal. Computational studies on five datasets, of which two are synthetically generated, illustrate clear advantages of our mixture-of-experts method for high-dimensional inputs, outperforming competitor models both in terms of validation metrics and visual inspection.
引用
收藏
相关论文
共 50 条
  • [1] A similarity-based Bayesian mixture-of-experts model
    Zhang, Tianfang
    Bokrantz, Rasmus
    Olsson, Jimmy
    [J]. STATISTICS AND COMPUTING, 2023, 33 (04)
  • [2] Mixture-of-Experts Variational Autoencoder for clustering and generating from similarity-based representations on single cell data
    Kopf, Andreas
    Fortuin, Vincent
    Somnath, Vignesh Ram
    Claassen, Manfred
    [J]. PLOS COMPUTATIONAL BIOLOGY, 2021, 17 (06)
  • [3] MoDE: A Mixture-of-Experts Model with Mutual Distillation among the Experts
    Xie, Zhitian
    Zhang, Yinger
    Zhuang, Chenyi
    Shi, Qitao
    Liu, Zhining
    Gu, Jinjie
    Zhang, Guannan
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 16067 - 16075
  • [4] Spatial Mixture-of-Experts
    Dryden, Nikoli
    Hoefler, Torsten
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [5] A Mixture-of-Experts Model for Antonym-Synonym Discrimination
    Xie, Zhipeng
    Zeng, Nan
    [J]. ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 558 - 564
  • [6] SPEECHMOE2: MIXTURE-OF-EXPERTS MODEL WITH IMPROVED ROUTING
    You, Zhao
    Feng, Shulin
    Su, Dan
    Yu, Dong
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7217 - 7221
  • [7] Parsimonious mixture-of-experts based on mean mixture of multivariate normal distributions
    Sepahdar, Afsaneh
    Madadi, Mohsen
    Balakrishnan, Narayanaswamy
    Jamalizadeh, Ahad
    [J]. STAT, 2022, 11 (01):
  • [8] Bayesian shrinkage in mixture-of-experts models: identifying robust determinants of class membership
    Gregor Zens
    [J]. Advances in Data Analysis and Classification, 2019, 13 : 1019 - 1051
  • [10] Asymptotic properties of mixture-of-experts models
    Olteanu, M.
    Rynkiewicz, J.
    [J]. NEUROCOMPUTING, 2011, 74 (09) : 1444 - 1449