Mutual, information, metric entropy and cumulative relative entropy risk

被引:0
|
作者
Haussler, D [1 ]
Opper, M
机构
[1] Univ Calif Santa Cruz, Santa Cruz, CA 95064 USA
[2] Univ Wurzburg, Wurzburg, Germany
来源
ANNALS OF STATISTICS | 1997年 / 25卷 / 06期
关键词
mutual information; Hellinger distance; relative entropy; metric entropy; minimax risk; Bayes risk; density estimation; Kullback-Leibler distance;
D O I
暂无
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Assume (P-theta: theta epsilon Theta) is a set of probability distributions with a common dominating measure on a complete separable metric space Y. A state theta* epsilon Theta is chosen by Nature. A statistician obtains n independent observations Y-1,...,Y-n from Y distributed according to P-theta.. For each time t between 1 and n, based on the observations Y-1,...,Yt-1, the statistician produces an estimated distribution (P) over cap(t) for P-theta* and suffers a loss L(P-theta., (P) over cap(t)). The cumulative risk for the statistician is the average total loss up to time n. Of special interest in information theory, data compression, mathematical finance, computational learning theory and statistical mechanics is the special case when the loss L(P-theta*, (P) over cap(t)) is the relative entropy between the true distribution Pg and the estimated distribution IS,. Here the cumulative Bayes risk from time 1 to n is the mutual information between the random parameter Theta* and the observations Y-1,...,Y-n. New bounds on this mutual information are given in terms of the Laplace transform of the Hellinger distance between pairs of distributions indexed by parameters in Theta. From these, bounds on the cumulative minimax risk are given in terms of the metric entropy of Theta with respect to the Hellinger distance. The assumptions required for these bounds are very general and do not depend on the choice of the dominating measure. They apply to both finite- and infinite-dimensional Theta. They apply in some cases where Y is infinite dimensional, in some cases where Y is not compact, in some cases where the distributions are not smooth and in some parametric cases where asymptotic normality of the posterior distribution fails.
引用
收藏
页码:2451 / 2492
页数:42
相关论文
共 50 条
  • [1] Belavkin-Staszewski Relative Entropy, Conditional Entropy, and Mutual Information
    Zhai, Yuan
    Yang, Bo
    Xi, Zhengjun
    ENTROPY, 2022, 24 (06)
  • [2] Mutual Information and Relative Entropy of Sequential Effect Algebras
    汪加梅
    武俊德
    Cho Minhyung
    Communications in Theoretical Physics, 2010, 54 (08) : 215 - 218
  • [3] Mutual Information, Relative Entropy, and Estimation in the Poisson Channel
    Atar, Rami
    Weissman, Tsachy
    2011 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY PROCEEDINGS (ISIT), 2011, : 708 - 712
  • [4] IT Formulae for Gamma Target: Mutual Information and Relative Entropy
    Arras, Benjamin
    Swan, Yvik
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2018, 64 (02) : 1083 - 1091
  • [5] Mutual Information, Relative Entropy, and Estimation in the Poisson Channel
    Atar, Rami
    Weissman, Tsachy
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2012, 58 (03) : 1302 - 1318
  • [6] Mutual Information and Relative Entropy of Sequential Effect Algebras
    Wang Jia-Mei
    Wu Jun-De
    Minhyung, Cho
    COMMUNICATIONS IN THEORETICAL PHYSICS, 2010, 54 (02) : 215 - 218
  • [7] Image fusion metric based on mutual information and Tsallis entropy
    Cvejic, N.
    Canagarajah, C. N.
    Bull, D. R.
    ELECTRONICS LETTERS, 2006, 42 (11) : 626 - 627
  • [8] Estimation of entropy and mutual information
    Paninski, L
    NEURAL COMPUTATION, 2003, 15 (06) : 1191 - 1253
  • [9] Mutual Information Is Copula Entropy
    马健
    孙增圻
    Tsinghua Science and Technology, 2011, 16 (01) : 51 - 54
  • [10] Mutual Information and Relative Entropy over the Binomial and Negative Binomial Channels
    Taborda, Camilo G.
    Perez-Cruz, Fernando
    2012 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY PROCEEDINGS (ISIT), 2012, : 696 - 700