Shannon's inequality for the Rényi entropy and an application to the uncertainty principle (Q2149498): Difference between revisions

From MaRDI portal
Added link to MaRDI item.
Import240304020342 (talk | contribs)
Set profile property.
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank

Revision as of 07:03, 5 March 2024

scientific article
Language Label Description Also known as
English
Shannon's inequality for the Rényi entropy and an application to the uncertainty principle
scientific article

    Statements

    Shannon's inequality for the Rényi entropy and an application to the uncertainty principle (English)
    0 references
    0 references
    29 June 2022
    0 references
    For a nonnegative function \(f\in \mathbb{L}^{1}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), the quantity defined by \[ h[f]= - \int_{\mathbb{R}^{n}}f(x)\log (f(x))dx \] is called the Shannon entropy. \textit{C. E. Shannon} himself proved [Bell Syst. Tech. J. 27, 379--423, 623--656 (1948; Zbl 1154.94303)] that the Shannon entropy is bounded by the second moment of the function. More precisely, if \(f\in \mathbb{L}^{2}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), then we have \[ -\int_{\mathbb{R}^{n}}f(x)\log(f(x))dx\leq \frac{n}{2}\log \left(\frac{2\pi e}{n}\int_{\mathbb{R}^{n}}|x|^{2}f(x)dx\right), \] where the constant \(\dfrac{2\pi e}{n}\) is optimal. For \(0 <\alpha< +\infty\) with \(\alpha \neq 1\), the Rényi entropy \(h_{\alpha}\) is defined by \[ h_{\alpha}[f]= \frac{1}{1-\alpha}\log \left(\int_{\mathbb{R}^{n}} f(x)^{\alpha}dx\right) \] for functions \(f\in \mathbb{L}^{1}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\). The main purpose of this paper is to prove the analog of the above inequality for Rényi entropies. More exactly, the author proves that if \(\alpha>0\) with \(\alpha\neq 1\) and \[ b> \begin{cases} 0, & \text{ if }\alpha>1\\ n\left(\frac{1}{\alpha}-1\right), & \text{ if }0< \alpha <1, \end{cases} \] then for all nonnegative function \(f\in \mathbb{L}^{1}_{b}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), there exists a constant \(C_{b}\) such that \[ \frac{1}{1-\alpha}\log \left(\int_{\mathbb{R}^{n}} f(x)^{\alpha}dx\right) \leq \frac{n}{b}\log \left(C_{b}\int_{\mathbb{R}^{n}}|x|^{b}f(x)dx\right). \] Further, the constant \(C_{b}\) is explicitly given and proved that it is optimal.
    0 references
    Shannon entropy
    0 references
    Shannon inequality
    0 references
    Rényi entropy
    0 references
    uncertainty principle
    0 references

    Identifiers