Shannon's inequality for the Rényi entropy and an application to the uncertainty principle (Q2149498): Difference between revisions

From MaRDI portal
Import240304020342 (talk | contribs)
Set profile property.
Set OpenAlex properties.
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1016/j.jfa.2022.109566 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W4281552283 / rank
 
Normal rank

Revision as of 20:08, 19 March 2024

scientific article
Language Label Description Also known as
English
Shannon's inequality for the Rényi entropy and an application to the uncertainty principle
scientific article

    Statements

    Shannon's inequality for the Rényi entropy and an application to the uncertainty principle (English)
    0 references
    0 references
    29 June 2022
    0 references
    For a nonnegative function \(f\in \mathbb{L}^{1}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), the quantity defined by \[ h[f]= - \int_{\mathbb{R}^{n}}f(x)\log (f(x))dx \] is called the Shannon entropy. \textit{C. E. Shannon} himself proved [Bell Syst. Tech. J. 27, 379--423, 623--656 (1948; Zbl 1154.94303)] that the Shannon entropy is bounded by the second moment of the function. More precisely, if \(f\in \mathbb{L}^{2}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), then we have \[ -\int_{\mathbb{R}^{n}}f(x)\log(f(x))dx\leq \frac{n}{2}\log \left(\frac{2\pi e}{n}\int_{\mathbb{R}^{n}}|x|^{2}f(x)dx\right), \] where the constant \(\dfrac{2\pi e}{n}\) is optimal. For \(0 <\alpha< +\infty\) with \(\alpha \neq 1\), the Rényi entropy \(h_{\alpha}\) is defined by \[ h_{\alpha}[f]= \frac{1}{1-\alpha}\log \left(\int_{\mathbb{R}^{n}} f(x)^{\alpha}dx\right) \] for functions \(f\in \mathbb{L}^{1}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\). The main purpose of this paper is to prove the analog of the above inequality for Rényi entropies. More exactly, the author proves that if \(\alpha>0\) with \(\alpha\neq 1\) and \[ b> \begin{cases} 0, & \text{ if }\alpha>1\\ n\left(\frac{1}{\alpha}-1\right), & \text{ if }0< \alpha <1, \end{cases} \] then for all nonnegative function \(f\in \mathbb{L}^{1}_{b}(\mathbb{R}^{n})\) with \(\left\| f\right\|_{1}=1\), there exists a constant \(C_{b}\) such that \[ \frac{1}{1-\alpha}\log \left(\int_{\mathbb{R}^{n}} f(x)^{\alpha}dx\right) \leq \frac{n}{b}\log \left(C_{b}\int_{\mathbb{R}^{n}}|x|^{b}f(x)dx\right). \] Further, the constant \(C_{b}\) is explicitly given and proved that it is optimal.
    0 references
    Shannon entropy
    0 references
    Shannon inequality
    0 references
    Rényi entropy
    0 references
    uncertainty principle
    0 references

    Identifiers