Examples of the application of nonparametric information geometry to statistical physics (Q280599): Difference between revisions

From MaRDI portal
Import240304020342 (talk | contribs)
Set profile property.
Set OpenAlex properties.
Property / OpenAlex ID
 
Property / OpenAlex ID: W3102391787 / rank
 
Normal rank

Revision as of 20:20, 19 March 2024

scientific article
Language Label Description Also known as
English
Examples of the application of nonparametric information geometry to statistical physics
scientific article

    Statements

    Examples of the application of nonparametric information geometry to statistical physics (English)
    0 references
    0 references
    0 references
    10 May 2016
    0 references
    The departure point of the present review paper is the monograph of \textit{S.-i. Amari} and \textit{H. Nagaoka} [Methods of information geometry. Providence, RI: American Mathematical Society; Oxford: Oxford University Press (2000; Zbl 0960.62005)] where the fundamental treatement of differential connections has been established for finite dimensional coordinate systems. Developments following A. P. Dawid's comments on [\textit{B. Efron}, Ann. Stat. 3, 1189--1242 (1975; Zbl 0321.62013)] enable a non-parametric extension of the Amari-Nagaoka theory where positive probability densities of a measure space were shown to be a differentiable manifold modeled on Orlicz banach spaces. A major purpose of the paper is to show that the pertinent mathematical formalism provides a rigorous way to interpret the Boltzmann equation as an evolution equation on the statistical manifold. Applications of Orlicz spaces in statistical physics have been previously exposed by \textit{W. A. Majewski} and \textit{L. E. Labuschagne} [Ann. Henri Poincaré 15, No. 6, 1197--1221 (2014; Zbl 1295.82019)]. The latter framework is presently worked out for a space-homogeneous Boltzmann operator with a suitable angular collision kernel. The non-parametric setting is also used to discuss a number of problems like e.g statistical physics of machine learning (in particular the black-box optimization), Kullback-Leibler divergence and Boltzmann-Gibbs entropy.
    0 references
    information geometry
    0 references
    exponential manifold
    0 references
    differential (statistical) connections
    0 references
    Boltzmann-Gibbs entropy
    0 references
    Boltzmann equation
    0 references
    Orlicz spaces
    0 references
    non-parametric statistical manifolds
    0 references
    Kullback-Leibler divergence
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references