Estimation of KL Divergence: Optimal Minimax Rate

被引:45
|
作者
Bu, Yuheng [1 ]
Zou, Shaofeng [1 ]
Liang, Yingbin [2 ,3 ]
Veeravalli, Venugopal V. [1 ]
机构
[1] Univ Illinois, ECE Dept, Coordinated Sci Lab, Urbana, IL 61801 USA
[2] Syracuse Univ, Dept Elect Engn & Comp Sci, Syracuse, NY 13244 USA
[3] Ohio State Univ, Dept Elect & Comp Engn, Columbus, OH 43210 USA
基金
美国国家科学基金会;
关键词
Functional approximation; mean squared error; minimax lower bound; plug-in estimator; polynomial approximation; FUNCTIONALS; ENTROPY;
D O I
10.1109/TIT.2018.2805844
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The problem of estimating the Kullback-Leibler divergence D(P parallel to Q) between two unknown distributions P and Q is studied, under the assumption that the alphabet size k of the distributions can scale to infinity. The estimation is based on m independent samples drawn from P and n independent samples drawn from Q. It is first shown that there does not exist any consistent estimator that guarantees asymptotically small worst case quadratic risk over the set of all pairs of distributions. A restricted set that contains pairs of distributions, with density ratio bounded by a function f (k) is further considered. An augmented plug-in estimator is proposed, and its worst case quadratic risk is shown to be within a constant factor of ((k/m) + (k f(k)/n))(2) + (log(2)f (k)/m) + (f (k)/n), if m and n exceed a constant factor of k and k f (k), respectively. Moreover, the minimax quadratic risk is characterized to be within a constant factor of ((k/(mlog k))+(k f (k)/(n log k)))(2) + (log(2)f (k)/m)+(f (k)/n), if m and n exceed a constant factor of k/log(k) and k f (k)/log k, respectively. The lower bound on the minimax quadratic risk is characterized by employing a generalized Le Cam's method. A minimax optimal estimator is then constructed by employing both the polynomial approximation and the plug-in approaches.
引用
收藏
页码:2648 / 2674
页数:27
相关论文
共 50 条
  • [1] Minimax Rate-optimal Estimation of KL Divergence between Discrete Distributions
    Han, Yanjun
    Jiao, Jiantao
    Weissman, Tsachy
    [J]. PROCEEDINGS OF 2016 INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY AND ITS APPLICATIONS (ISITA 2016), 2016, : 256 - 260
  • [2] Minimax Optimal Estimation of KL Divergence for Continuous Distributions
    Zhao, Puning
    Lai, Lifeng
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2020, 66 (12) : 7787 - 7811
  • [3] MINIMAX ESTIMATION WITH DIVERGENCE LOSS FUNCTION
    KASHYAP, RL
    [J]. INFORMATION SCIENCES, 1974, 7 (3-4) : 341 - 364
  • [4] Minimax Optimal Rate for Parameter Estimation in Multivariate Deviated Models
    Do, Dat
    Nguyen, Huy
    Nguyen, Khai
    Ho, Nhat
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Minimax Optimal Additive Functional Estimation with Discrete Distribution: Slow Divergence Speed Case
    Fukuchi, Kazuto
    Sakuma, Jun
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 1041 - 1045
  • [6] Minimax estimation of norms of a probability density: II. Rate-optimal estimation procedures
    Goldenshluger, Alexander
    Lepski, Oleg, V
    [J]. BERNOULLI, 2022, 28 (02) : 1155 - 1178
  • [7] Profile Maximum Likelihood is Optimal for Estimating KL Divergence
    Acharya, Jayadev
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 1400 - 1404
  • [8] Minimax quantum state estimation under Bregman divergence
    Quadeer, Maria
    Tomamichel, Marco
    Ferrie, Christopher
    [J]. QUANTUM, 2019, 3
  • [9] Minimax-Optimal Location Estimation
    Gupta, Shivam
    Lee, Jasper C. H.
    Price, Eric
    Valiant, Paul
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Estimation of KL Divergence Between Large-Alphabet Distributions
    Bu, Yuheng
    Zou, Shaofeng
    Liang, Yingbin
    Veeravalli, Venugopal V.
    [J]. 2016 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, 2016, : 1118 - 1122