Twenty Years of Mixture of Experts

被引:330
|
作者
Yuksel, Seniha Esen [1 ]
Wilson, Joseph N. [1 ]
Gader, Paul D. [1 ]
机构
[1] Univ Florida, Dept Comp & Informat Sci & Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
Applications; Bayesian; classification; comparison; hierarchical mixture of experts (HME); mixture of Gaussian process experts; regression; statistical properties; survey; variational; TIME-SERIES PREDICTION; INDEPENDENT FACE RECOGNITION; SUPPORT VECTOR MACHINES; OF-EXPERTS; HIERARCHICAL MIXTURES; NEURAL-NETWORKS; EM ALGORITHM; ASYMPTOTIC NORMALITY; MAXIMUM-LIKELIHOOD; BAYESIAN-INFERENCE;
D O I
10.1109/TNNLS.2012.2200299
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we provide a comprehensive survey of the mixture of experts (ME). We discuss the fundamental models for regression and classification and also their training with the expectation-maximization algorithm. We follow the discussion with improvements to the ME model and focus particularly on the mixtures of Gaussian process experts. We provide a review of the literature for other training methods, such as the alternative localized ME training, and cover the variational learning of ME in detail. In addition, we describe the model selection literature which encompasses finding the optimum number of experts, as well as the depth of the tree. We present the advances in ME in the classification area and present some issues concerning the classification model. We list the statistical properties of ME, discuss how the model has been modified over the years, compare ME to some popular algorithms, and list several applications. We conclude our survey with future directions and provide a list of publicly available datasets and a list of publicly available software that implement ME. Finally, we provide examples for regression and classification. We believe that the study described in this paper will provide quick access to the relevant literature for researchers and practitioners who would like to improve or use ME, and that it will stimulate further studies in ME.
引用
收藏
页码:1177 / 1193
页数:17
相关论文
共 50 条
  • [1] Mixture of vector experts
    Henderson, M
    Shawe-Taylor, J
    Zerovnik, J
    ALGORITHMIC LEARNING THEORY, 2005, 3734 : 386 - 398
  • [2] Twenty years already, barely twenty years
    Hutten-Czapski, Peter
    CANADIAN JOURNAL OF RURAL MEDICINE, 2016, 21 (02) : 36 - 36
  • [3] Constrained Mixture Models of Soft Tissue Growth and Remodeling - Twenty Years After
    Humphrey, J. D.
    JOURNAL OF ELASTICITY, 2021, 145 (1-2) : 49 - 75
  • [4] Constrained Mixture Models of Soft Tissue Growth and Remodeling – Twenty Years After
    J. D. Humphrey
    Journal of Elasticity, 2021, 145 : 49 - 75
  • [5] Twenty articles for twenty years
    Waddington, Gordon
    JOURNAL OF SCIENCE AND MEDICINE IN SPORT, 2017, 20 (12) : 1132 - 1132
  • [6] Twenty Sox, twenty years
    Harley, Vincent
    Lefebvre, Veronique
    INTERNATIONAL JOURNAL OF BIOCHEMISTRY & CELL BIOLOGY, 2010, 42 (03): : 376 - 377
  • [7] Latent Mixture of Discriminative Experts
    Ozkan, Derya
    Morency, Louis-Philippe
    IEEE TRANSACTIONS ON MULTIMEDIA, 2013, 15 (02) : 326 - 338
  • [8] Hierarchical Routing Mixture of Experts
    Zhao, Wenbo
    Gao, Yang
    Memon, Shahan Ali
    Raj, Bhiksha
    Singh, Rita
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7900 - 7906
  • [9] Mixture of experts: a literature survey
    Masoudnia, Saeed
    Ebrahimpour, Reza
    ARTIFICIAL INTELLIGENCE REVIEW, 2014, 42 (02) : 275 - 293
  • [10] Mixture of Experts with Genetic Algorithms
    Cleofas, Laura
    Maria Valdovinos, Rosa
    Juarez, C.
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, 2009, 61 : 331 - 338