Strong identifiability and optimal minimax rates for finite mixture estimation (Q1991679): Difference between revisions

From MaRDI portal
Added link to MaRDI item.
Import240304020342 (talk | contribs)
Set profile property.
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank

Revision as of 05:29, 5 March 2024

scientific article
Language Label Description Also known as
English
Strong identifiability and optimal minimax rates for finite mixture estimation
scientific article

    Statements

    Strong identifiability and optimal minimax rates for finite mixture estimation (English)
    0 references
    0 references
    0 references
    30 October 2018
    0 references
    Excerpt from abstract and paper: ``We study the rates of estimation of finite mixing distributions, that is, the parameters of the mixture. We prove that under some regularity and strong identifiability conditions, around a given mixing distribution with $m_0$ components, the optimal local minimax rate of estimation of a mixing distribution with $m$ components is $n^{-1/(4(m-m_0)+2)}$. This corrects a previous paper by \textit{J. Chen} [Ann. Stat. 23, No. 1, 221--233 (1995; Zbl 0821.62023)]. [\dots] The rate gets worse with more components, which is consistent with the behaviour when there are infinitely many components, such as deconvolution. [...] In addition, the optimal local minimax rate and the optimal pointwise rate of estimation everywhere are not the same. This discrepancy is unusual in statistics, and probably the reason why the $n^{-1/4}$ rate went unchallenged for twenty years. Specifically, if instead of comparing all pairs of mixtures in a ball, we allow only one mixture in it, we get (21) which corrects Lemma 2 of Chen [loc. cit.]. As a consequence, Theorem 2 of Chen is valid by dropping uniformity: for any fixed mixing distribution say $G$, the estimator considered there will converge at rate $n^{-1/4}$, but with a multiplicative constant that depends on $G$. It then becomes a statement on the optimal pointwise rate of estimation everywhere, and can even be strengthened to $n^{-1/2}$.''
    0 references
    0 references
    local asymptotic normality
    0 references
    convergence of experiments
    0 references
    maximum likelihood estimate
    0 references
    Wasserstein metric
    0 references
    mixing distribution
    0 references
    mixture model
    0 references
    rate of convergence
    0 references
    strong identifiability
    0 references
    pointwise rate
    0 references
    superefficiency
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references