STATISTICAL PROBLEM CLASSES AND THEIR LINKS TO INFORMATION THEORY

被引:3
|
作者
Clarke, Bertrand [1 ,2 ]
Clarke, Jennifer [3 ]
Yu, Chi Wai [4 ]
机构
[1] Univ Miami, Dept Med, Ctr Computat Sci, Miami, FL 33136 USA
[2] Univ Miami, Dept Epidemiol & Publ Hlth, Ctr Computat Sci, Miami, FL 33136 USA
[3] Univ Miami, Dept Epidemiol & Publ Hlth, Miami, FL USA
[4] Hong Kong Univ Sci & Technol, Dept Math, Hong Kong, Hong Kong, Peoples R China
关键词
Bayesian; Codelength; Entropy; Information theory; M-closed; M-complete; M-open; Mutual information; Model selection; Prediction; Relative entropy; Rate distortion; C52; C53; MODEL; SELECTION; LIKELIHOOD; COMPUTATION; CONSISTENCY; COMPLEXITY; DIMENSION; CRITERION; CAPACITY; ENTROPY;
D O I
10.1080/07474938.2013.807190
中图分类号
F [经济];
学科分类号
02 ;
摘要
We begin by recalling the tripartite division of statistical problems into three classes, M-closed, M-complete, and M-open and then reviewing the key ideas of introductory Shannon theory. Focusing on the related but distinct goals of model selection and prediction, we argue that different techniques for these two goals are appropriate for the three different problem classes. For M-closed problems we give relative entropy justification that the Bayes information criterion (BIC) is appropriate for model selection and that the Bayes model average is information optimal for prediction. For M-complete problems, we discuss the principle of maximum entropy and a way to use the rate distortion function to bypass the inaccessibility of the true distribution. For prediction in the M-complete class, there is little work done on information based model averaging so we discuss the Akaike information criterion (AIC) and its properties and variants. For the M-open class, we argue that essentially only predictive criteria are suitable. Thus, as an analog to model selection, we present the key ideas of prediction along a string under a codelength criterion and propose a general form of this criterion. Since little work appears to have been done on information methods for general prediction in the M-open class of problems, we mention the field of information theoretic learning in certain general function spaces.
引用
收藏
页码:337 / 371
页数:35
相关论文
共 50 条