Continuity and characterization of Shannon-Wiener information measure for continuous probability distributions (Q2393890): Difference between revisions

From MaRDI portal
Importer (talk | contribs)
Created a new Item
 
Set OpenAlex properties.
 
(3 intermediate revisions by 3 users not shown)
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / cites work
 
Property / cites work: A Mathematical Theory of Communication / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5801644 / rank
 
Normal rank
Property / cites work
 
Property / cites work: A note on the entropy of a continuous distribution / rank
 
Normal rank
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1007/bf01737401 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2037790254 / rank
 
Normal rank
links / mardi / namelinks / mardi / name
 

Latest revision as of 10:56, 30 July 2024

scientific article
Language Label Description Also known as
English
Continuity and characterization of Shannon-Wiener information measure for continuous probability distributions
scientific article

    Statements

    Continuity and characterization of Shannon-Wiener information measure for continuous probability distributions (English)
    0 references
    0 references
    0 references
    1959
    0 references
    Let \((R, S, m)\) be a \(\sigma\)-finite measure space. \(V\) denotes the class of all absolutely continuous probability distributions \((X, X', X_n,\dots,\) say) with the densities \((dX/dm=p(x)\), \(dX'/dm = p'(x)\), \(dX_n/dm = p_n(x),\dots,\) say). The uniform metric in \(V\) is defined by \(d(X,X') = \text{ess.}\sup | p(x) - p'(x)|\), and the entropy of each \(X\in V\) is defined by \(H(X) = \int p(x) \log p(x) \,dm\). The author proves that: (1) When \(m(R)< \infty\), \(d(X_n,X)\to 0\) implies \(H(X_n) \to H(X)\) \((n\to\infty)\). (2) When \(m(R) = \infty\), if \(X\in V\) is mutually absolutely continuous with respect to \(m\), \(d(X_n,X)\to 0\) and \(\text{ess.} \sup_{p(x)>0} | 1 - p_n(x)/p(x)| \to 0\) \((n\to\infty)\) then \(H(X_n)\to H(X)\). A characterization theorem of the entropy \(H(X)\) of any continuously valued finite dimensional random variable \(X\) has been previously proved by \textit{H. Hatori} [Kodai Math. Semin. Rep. 10, 172--176 (1958; Zbl 0087.33201)]. The present author treats a similar characterization of \(H(X)\) in the abstract \(\sigma\)-finite measure space \((R, S, m)\), which is based upon the preceding considerations (1), (2) and two additional properties corresponding to the entropy of the uniform distribution and the conditional entropy.
    0 references
    0 references
    Shannon-Wiener information measure
    0 references
    continuous probability distributions
    0 references
    0 references