Abstract: At a superficial level, the idea of maximum likelihood must be prehistoric: early hunters and gatherers may not have used the words ``method of maximum likelihood to describe their choice of where and how to hunt and gather, but it is hard to believe they would have been surprised if their method had been described in those terms. It seems a simple, even unassailable idea: Who would rise to argue in favor of a method of minimum likelihood, or even mediocre likelihood? And yet the mathematical history of the topic shows this ``simple idea is really anything but simple. Joseph Louis Lagrange, Daniel Bernoulli, Leonard Euler, Pierre Simon Laplace and Carl Friedrich Gauss are only some of those who explored the topic, not always in ways we would sanction today. In this article, that history is reviewed from back well before Fisher to the time of Lucien Le Cam's dissertation. In the process Fisher's unpublished 1930 characterization of conditions for the consistency and efficiency of maximum likelihood estimates is presented, and the mathematical basis of his three proofs discussed. In particular, Fisher's derivation of the information inequality is seen to be derived from his work on the analysis of variance, and his later approach via estimating functions was derived from Euler's Relation for homogeneous functions. The reaction to Fisher's work is reviewed, and some lessons drawn.
Recommendations
Cites work
- scientific article; zbMATH DE number 3117956 (Why is no real title available?)
- scientific article; zbMATH DE number 3117973 (Why is no real title available?)
- scientific article; zbMATH DE number 4072104 (Why is no real title available?)
- scientific article; zbMATH DE number 4089504 (Why is no real title available?)
- scientific article; zbMATH DE number 1381973 (Why is no real title available?)
- scientific article; zbMATH DE number 3771560 (Why is no real title available?)
- scientific article; zbMATH DE number 48091 (Why is no real title available?)
- scientific article; zbMATH DE number 3626409 (Why is no real title available?)
- scientific article; zbMATH DE number 1055955 (Why is no real title available?)
- scientific article; zbMATH DE number 1064661 (Why is no real title available?)
- scientific article; zbMATH DE number 1064665 (Why is no real title available?)
- scientific article; zbMATH DE number 1138977 (Why is no real title available?)
- scientific article; zbMATH DE number 3448510 (Why is no real title available?)
- scientific article; zbMATH DE number 2108956 (Why is no real title available?)
- scientific article; zbMATH DE number 3224089 (Why is no real title available?)
- scientific article; zbMATH DE number 3251886 (Why is no real title available?)
- scientific article; zbMATH DE number 3029381 (Why is no real title available?)
- scientific article; zbMATH DE number 3028455 (Why is no real title available?)
- scientific article; zbMATH DE number 3061356 (Why is no real title available?)
- scientific article; zbMATH DE number 3097675 (Why is no real title available?)
- scientific article; zbMATH DE number 3103824 (Why is no real title available?)
- A Survey of Maximum Likelihood Estimation
- A history of parametric statistical inference from Bernoulli to Fisher, 1713--1935
- An invariant form for the prior probability in estimation problems
- Ancillary history
- Assessing the accuracy of the maximum likelihood estimator: Observed versus expected Fisher information
- Asymptotic Statistics
- Biometrika centenary: Sample surveys
- Consistency of the Maximum Likelihood Estimator in the Presence of Infinitely Many Incidental Parameters
- Consistent Estimates Based on Partially Consistent Observations
- Defining the curvature of a statistical problem (with applications to second order efficiency)
- F. Y. Edgeworth and R. A. Fisher on the efficiency of maximum likelihood estimation
- Fisher in 1921
- Harold Hotelling 1895--1973
- J. H. Lambert's work on probability
- Karl Pearson's theoretical errors and the advances they inspired
- Laplace's 1774 memoir on inverse probability
- Maximum Likelihood: An Introduction
- Maximum likelihood and decision theory
- Note on the Consistency of the Maximum Likelihood Estimate
- On Fisher's Bound for Asymptotic Variances
- On rereading R. A. Fisher
- Outline of a Theory of Statistical Estimation Based on the Classical Theory of Probability
- Principles of Statistical Inference
- Probability and Statistics
- R. A. Fisher and the fiducial argument
- R. A. Fisher and the making of maximum likelihood 1912--1922
- R. A. Fisher in the 21st century. Invited paper presented at the 1996 R. A. Fisher lecture. (With comments).
- R. A. Fisher: an appreciation
- Statistical Estimation
- THE STATISTICAL UTILIZATION OF MULTIPLE MEASUREMENTS
- Tests of Statistical Hypotheses Concerning Several Parameters When the Number of Observations is Large
- The History of Likelihood
- The Impact of R. A. Fisher on Statistics
- The geometry of asymptotic inference. With comments and a rejoinder by the author
- The geometry of exponential families
- Thiele: Pioneer in Statistics
- Three early papers on efficient parametric estimation
- What did Fisher mean by inverse probability in 1912--1922?
Cited in
(26)- Big Bayes stories -- foreword
- Monte Carlo gradient estimation in machine learning
- Lagrange and probability theory
- Interview with Myfanwy E. Evans: entanglements on and models of periodic minimal surfaces
- Half a century of information geometry, part 1
- Clustering in Hilbert's projective geometry: the case studies of the probability simplex and the elliptope of correlation matrices
- R. A. Fisher and the making of maximum likelihood 1912--1922
- Suppression of metastasis by primary tumor and acceleration of metastasis following primary tumor resection: A natural law?
- Karl Pearson's theoretical errors and the advances they inspired
- Extreme value distributions: an overview of estimation and simulation
- Lectures on Entropy. I: Information-Theoretic Notions
- Maximum likelihood characterization of distributions
- On the history of maximum likelihood in relation to inverse probability and least squares.
- Superefficiency from the vantage point of computability
- Stein 1956: Efficient nonparametric testing and estimation
- What did Fisher mean by inverse probability in 1912--1922?
- Doob at Lyon: Bringing Martingales Back to France
- Model-based geostatistics from a Bayesian perspective: investigating area-to-point Kriging with small data sets
- On closed-form expressions for the Fisher-Rao distance
- On Fitting Probability Distribution to Univariate Grouped Actuarial Data with Both Group Mean and Relative Frequencies
- Analytic posteriors for Pearson's correlation coefficient
- A conversation with Stephen M. Stigler
- Maximum Likelihood Estimation and Graph Matching in Errorfully Observed Networks
- Testing that a local optimum of the likelihood is globally optimum using reparameterized embeddings. Applications to wavefront sensing
- Mere Renovation is Too Little Too Late: We Need to Rethink our Undergraduate Curriculum from the Ground Up
- The double Gaussian approximation for high frequency data
This page was built for publication: The epic story of maximum likelihood
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q449800)