Entropy estimation via uniformization

From MaRDI portal
Publication:6136099

DOI10.1016/J.ARTINT.2023.103954arXiv2304.09700OpenAlexW4380356568MaRDI QIDQ6136099FDOQ6136099

Jinglai Li, Ziqiao Ao

Publication date: 28 August 2023

Published in: Artificial Intelligence (Search for Journal in Brave)

Abstract: Entropy estimation is of practical importance in information theory and statistical science. Many existing entropy estimators suffer from fast growing estimation bias with respect to dimensionality, rendering them unsuitable for high-dimensional problems. In this work we propose a transform-based method for high-dimensional entropy estimation, which consists of the following two main ingredients. First by modifying the k-NN based entropy estimator, we propose a new estimator which enjoys small estimation bias for samples that are close to a uniform distribution. Second we design a normalizing flow based mapping that pushes samples toward a uniform distribution, and the relation between the entropy of the original samples and the transformed ones is also derived. As a result the entropy of a given set of samples is estimated by first transforming them toward a uniform distribution and then applying the proposed estimator to the transformed samples. The performance of the proposed method is compared against several existing entropy estimators, with both mathematical examples and real-world applications.


Full work available at URL: https://arxiv.org/abs/2304.09700







Cites Work






This page was built for publication: Entropy estimation via uniformization

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6136099)