Bayesian classification using an entropy prior on mixture models

被引:0
|
作者
Center, JL [1 ]
机构
[1] Creat Res Corp, Andover, MA 01810 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In many classification problems, it is reasonable to base the analysis on a mixture model, A mixture model assumes that each sample is produced by first randomly selecting from a finite collection of data clusters and by then using the chosen cluster distribution to produce the class label and feature vector of the sample. If we know the set of model parameters, then when we observe a feature vector, we can predict the classification. When we do not know the parameters exactly, we must infer the model parameters from a training set of data samples. Taking the Bayesian approach, we want to determine the probability distribution for the parameters given the training data. Then when it comes time to predict the class label, given a feature vector, we integrate over the model parameter distribution. We argue that a good, objective choice for the prior distribution on the model parameters is based on the entropy of each mixture model. We show that this prior regularizes the model fit so that over-fitting the training data has no adverse effects.
引用
收藏
页码:42 / 70
页数:29
相关论文
共 50 条
  • [1] Bayesian estimation and classification with incomplete data using mixture models
    Zhang, JF
    Everson, R
    [J]. PROCEEDINGS OF THE 2004 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA'04), 2004, : 296 - 303
  • [2] Incorporation of prior knowledge into genetic association studies using Bayesian mixture models
    Fridley, Brooke
    Serie, Daniel
    White, Kristin
    Jenkins, Gregory
    Bamlet, William
    Goode, Ellen
    [J]. CANCER RESEARCH, 2009, 69
  • [3] Classification of molecular sequence data using Bayesian phylogenetic mixture models
    Loza-Reyes, E.
    Hum, M. A.
    Robinson, A.
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2014, 75 : 81 - 95
  • [4] Bayesian Learning for Classification using a Uniform Dirichlet Prior
    Rademacher, Paul
    Doroslovacki, Milos
    [J]. 2019 7TH IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (IEEE GLOBALSIP), 2019,
  • [5] Fast Search and Estimation of Bayesian Nonparametric Mixture Models Using a Classification Annealing EM Algorithm
    Karabatsos, George
    [J]. JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2021, 30 (01) : 236 - 247
  • [6] Bayesian Mixture Models for the Incorporation of Prior Knowledge to Inform Genetic Association Studies
    Fridley, Brooke L.
    Serie, Daniel
    Jenkins, Gregory
    White, Kristin
    Bamlet, William
    Potter, John D.
    Goode, Ellen L.
    [J]. GENETIC EPIDEMIOLOGY, 2010, 34 (05) : 418 - 426
  • [7] Improving Subpixel Classification by Incorporating Prior Information in Linear Mixture Models
    Kasetkasem, Teerasit
    Arora, Manoj K.
    Varshney, Pramod K.
    Areekul, Vutipong
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2011, 49 (03): : 1001 - 1013
  • [8] Entropy-type classification maximum likelihood algorithms for mixture models
    Chien-Yo Lai
    Miin-Shen Yang
    [J]. Soft Computing, 2011, 15 : 373 - 381
  • [9] Entropy-type classification maximum likelihood algorithms for mixture models
    Lai, Chien-Yo
    Yang, Miin-Shen
    [J]. SOFT COMPUTING, 2011, 15 (02) : 373 - 381
  • [10] Bayesian subgroup analysis in regression using mixture models
    Im, Yunju
    Tan, Aixin
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2021, 162