A Kullback-Leibler View of Maximum Entropy and Maximum Log-Probability Methods

被引:21
|
作者
Abbas, Ali E. [1 ]
Cadenbach, Andrea H. [2 ]
Salimi, Ehsan [3 ]
机构
[1] Univ Southern Calif, Ind & Syst Engn & Publ Policy, Los Angeles, CA 90089 USA
[2] Univ Missouri, Supply Chain & Analyt, St Louis, MO 63121 USA
[3] Univ Southern Calif, Ind & Syst Engn, Los Angeles, CA 90007 USA
来源
ENTROPY | 2017年 / 19卷 / 05期
基金
美国国家科学基金会;
关键词
entropy; minimum cross entropy; joint probability distribution; DISTRIBUTIONS; INFORMATION;
D O I
10.3390/e19050232
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Entropy methods enable a convenient general approach to providing a probability distribution with partial information. The minimum cross-entropy principle selects the distribution that minimizes the Kullback-Leibler divergence subject to the given constraints. This general principle encompasses a wide variety of distributions, and generalizes other methods that have been proposed independently. There remains, however, some confusion about the breadth of entropy methods in the literature. In particular, the asymmetry of the Kullback-Leibler divergence provides two important special cases when the target distribution is uniform: the maximum entropy method and the maximum log-probability method. This paper compares the performance of both methods under a variety of conditions. We also examine a generalized maximum log-probability method as a further demonstration of the generality of the entropy approach.
引用
收藏
页数:14
相关论文
共 50 条