Entropy estimation via uniformization
From MaRDI portal
Publication:6136099
DOI10.1016/J.ARTINT.2023.103954arXiv2304.09700OpenAlexW4380356568MaRDI QIDQ6136099FDOQ6136099
Publication date: 28 August 2023
Published in: Artificial Intelligence (Search for Journal in Brave)
Abstract: Entropy estimation is of practical importance in information theory and statistical science. Many existing entropy estimators suffer from fast growing estimation bias with respect to dimensionality, rendering them unsuitable for high-dimensional problems. In this work we propose a transform-based method for high-dimensional entropy estimation, which consists of the following two main ingredients. First by modifying the k-NN based entropy estimator, we propose a new estimator which enjoys small estimation bias for samples that are close to a uniform distribution. Second we design a normalizing flow based mapping that pushes samples toward a uniform distribution, and the relation between the entropy of the original samples and the transformed ones is also derived. As a result the entropy of a given set of samples is estimated by first transforming them toward a uniform distribution and then applying the proposed estimator to the transformed samples. The performance of the proposed method is compared against several existing entropy estimators, with both mathematical examples and real-world applications.
Full work available at URL: https://arxiv.org/abs/2304.09700
Statistical aspects of information-theoretic topics (62B10) Measures of information, entropy (94A17)
Cites Work
- A Mathematical Theory of Communication
- Sample estimate of the entropy of a random vector
- The jackknife estimate of variance
- Title not available (Why is that?)
- Combinatorics of partial derivatives
- Efficient multivariate entropy estimation via \(k\)-nearest neighbour distances
- Title not available (Why is that?)
- A new class of random vector entropy estimators and its applications in testing statistical hypotheses
- On the estimation of entropy
- Title not available (Why is that?)
- Title not available (Why is that?)
- Title not available (Why is that?)
- Estimation of entropy and other functionals of a multivariate density
- Maximum Entropy Sampling and Optimal Bayesian Experimental Design
- Minimum-entropy estimation in semi-parametric models
- Estimation of integral functionals of a density
- Density-free convergence properties of various estimators of entropy
- Title not available (Why is that?)
- Lectures on the nearest neighbor method
- Towards Bayesian experimental design for nonlinear models that require a large number of sampling times
- Demystifying Fixed <inline-formula> <tex-math notation="LaTeX">$k$ </tex-math> </inline-formula>-Nearest Neighbor Information Estimators
- Ensemble Estimators for Multivariate Entropy Estimation
- Optimal rates of entropy estimation over Lipschitz balls
- Geometric k-nearest neighbor estimation of entropy and mutual information
- SENSITIVITY ANALYSIS FOR STOCHASTIC SIMULATORS USING DIFFERENTIAL ENTROPY
This page was built for publication: Entropy estimation via uniformization
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6136099)