The relation between information theory and the differential geometry approach to statistics (Q1086923)

From MaRDI portal
scientific article
Language Label Description Also known as
English
The relation between information theory and the differential geometry approach to statistics
scientific article

    Statements

    The relation between information theory and the differential geometry approach to statistics (English)
    0 references
    1985
    0 references
    \textit{N. N. Chentsov} [Statistical decision rules and optimal inference. Transl. Math. Monogr. 53 (1982; Zbl 0484.62008)] has shown that the Riemannian metric on the probability simplex \(\sum x_ i=1\) defined by \((ds)^ 2=\sum (dx_ i)^ 2/x_ i\) has an invariance property under certain probabilistically natural mappings. No other Riemannian metric has the same property. The geometry associated with this metric is shown to lead almost automatically to measures of divergence between probability distributions which are associated with Kullback, Bhattacharyya, and Matusita. Certain vector fields are associated in a natural way with random variables. The integral curves of these vector fields yield the maximum entropy or minimum divergence estimates of probabilities. Some other consequences of this geometric view are also explored.
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    information theory
    0 references
    information measures
    0 references
    Fisher information matrix
    0 references
    Bhattacharyya distance
    0 references
    Kullback distance
    0 references
    Matusita distance
    0 references
    Riemannian metric
    0 references
    probability simplex
    0 references
    measures of divergence
    0 references
    vector fields
    0 references
    maximum entropy
    0 references
    minimum divergence estimates
    0 references
    0 references
    0 references
    0 references