Estimation of KL Divergence: Optimal Minimax Rate

From MaRDI portal
Publication:4569211

DOI10.1109/TIT.2018.2805844zbMATH Open1390.94614arXiv1607.02653OpenAlexW2963305780MaRDI QIDQ4569211FDOQ4569211


Authors: Yuheng Bu, Shaofeng Zou, Yingbin Liang, Venugopal V. Veeravalli Edit this on Wikidata


Publication date: 27 June 2018

Published in: IEEE Transactions on Information Theory (Search for Journal in Brave)

Abstract: The problem of estimating the Kullback-Leibler divergence D(P|Q) between two unknown distributions P and Q is studied, under the assumption that the alphabet size k of the distributions can scale to infinity. The estimation is based on m independent samples drawn from P and n independent samples drawn from Q. It is first shown that there does not exist any consistent estimator that guarantees asymptotically small worst-case quadratic risk over the set of all pairs of distributions. A restricted set that contains pairs of distributions, with density ratio bounded by a function f(k) is further considered. {An augmented plug-in estimator is proposed, and its worst-case quadratic risk is shown to be within a constant factor of (frackm+frackf(k)n)2+fraclog2f(k)m+fracf(k)n, if m and n exceed a constant factor of k and kf(k), respectively.} Moreover, the minimax quadratic risk is characterized to be within a constant factor of (frackmlogk+frackf(k)nlogk)2+fraclog2f(k)m+fracf(k)n, if m and n exceed a constant factor of k/log(k) and kf(k)/logk, respectively. The lower bound on the minimax quadratic risk is characterized by employing a generalized Le Cam's method. A minimax optimal estimator is then constructed by employing both the polynomial approximation and the plug-in approaches.


Full work available at URL: https://arxiv.org/abs/1607.02653




Recommendations




Cited In (5)





This page was built for publication: Estimation of KL Divergence: Optimal Minimax Rate

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4569211)