Shannon's inequality for the Rényi entropy and an application to the uncertainty principle (Q2149498)

From MaRDI portal
Revision as of 20:08, 19 March 2024 by Openalex240319060354 (talk | contribs) (Set OpenAlex properties.)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
scientific article
Language Label Description Also known as
English
Shannon's inequality for the Rényi entropy and an application to the uncertainty principle
scientific article

    Statements

    Shannon's inequality for the Rényi entropy and an application to the uncertainty principle (English)
    0 references
    0 references
    29 June 2022
    0 references
    For a nonnegative function \(f\in \mathbb{L}^{1}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), the quantity defined by \[ h[f]= - \int_{\mathbb{R}^{n}}f(x)\log (f(x))dx \] is called the Shannon entropy. \textit{C. E. Shannon} himself proved [Bell Syst. Tech. J. 27, 379--423, 623--656 (1948; Zbl 1154.94303)] that the Shannon entropy is bounded by the second moment of the function. More precisely, if \(f\in \mathbb{L}^{2}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), then we have \[ -\int_{\mathbb{R}^{n}}f(x)\log(f(x))dx\leq \frac{n}{2}\log \left(\frac{2\pi e}{n}\int_{\mathbb{R}^{n}}|x|^{2}f(x)dx\right), \] where the constant \(\dfrac{2\pi e}{n}\) is optimal. For \(0 <\alpha< +\infty\) with \(\alpha \neq 1\), the Rényi entropy \(h_{\alpha}\) is defined by \[ h_{\alpha}[f]= \frac{1}{1-\alpha}\log \left(\int_{\mathbb{R}^{n}} f(x)^{\alpha}dx\right) \] for functions \(f\in \mathbb{L}^{1}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\). The main purpose of this paper is to prove the analog of the above inequality for Rényi entropies. More exactly, the author proves that if \(\alpha>0\) with \(\alpha\neq 1\) and \[ b> \begin{cases} 0, & \text{ if }\alpha>1\\ n\left(\frac{1}{\alpha}-1\right), & \text{ if }0< \alpha <1, \end{cases} \] then for all nonnegative function \(f\in \mathbb{L}^{1}_{b}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), there exists a constant \(C_{b}\) such that \[ \frac{1}{1-\alpha}\log \left(\int_{\mathbb{R}^{n}} f(x)^{\alpha}dx\right) \leq \frac{n}{b}\log \left(C_{b}\int_{\mathbb{R}^{n}}|x|^{b}f(x)dx\right). \] Further, the constant \(C_{b}\) is explicitly given and proved that it is optimal.
    0 references
    0 references
    Shannon entropy
    0 references
    Shannon inequality
    0 references
    Rényi entropy
    0 references
    uncertainty principle
    0 references
    0 references