Construction of improved estimators in multiparameter estimation for continuous exponential families (Q792046)

From MaRDI portal





scientific article; zbMATH DE number 3852239
Language Label Description Also known as
default for all languages
No label defined
    English
    Construction of improved estimators in multiparameter estimation for continuous exponential families
    scientific article; zbMATH DE number 3852239

      Statements

      Construction of improved estimators in multiparameter estimation for continuous exponential families (English)
      0 references
      0 references
      0 references
      0 references
      1984
      0 references
      For a multivariate normal distribution \(N(\mu,\sigma^ 2I_ p)\), \textit{W. James} and \textit{C. Stein} [Estimation with quadratic loss. Proc. 4th Berkeley Sympos. math. statist. Probab. 1, 361-379 (1961)] proved the existence of an estimator \(\delta\) (x) for \(\mu\) which dominates the usual maximum likelihood (UMVUE, best invariant) estimator \(\delta^ 0(x)=x\). \(\delta\) is obtained by shrinking \(\delta^ 0\) towards 0. Similar questions have been investigated for continuous exponential families with a p-variate parameter \(\mu\) by \textit{J. Berger} [Ann. Stat. 8, 545-571 (1980; Zbl 0447.62008)] and by the first author and \textit{A. Parsian} [J. Multivariate Anal. 10, 551-564 (1980; Zbl 0453.62007)]. The paper at hand generalizes these results by constructing, for a given initial estimator \(\delta^ 0\), an improved estimator \(\delta\) that shrinks \(\delta^ 0\) towards a prechosen point \(m\in {\mathbb{R}}^ p\) or towards the (geometric or arithmetic) mean of the p components. Typically, \(\delta\) has the form \(\delta_ i(x)=\delta^ 0_ i(x)- \eta(x)h_ i(x_ i)\Phi_ i(x)\) with, e.g., \(\Phi_ i(x):=- (\tau(S)/S)(g_ i(x_ i)-m_ i)\), \(S:=\Sigma_ jd_ j| g_ j(x_ j)-m_ j|^{\beta}\) and suitable functions \(\tau\) (S), \(g_ i(x_ i)\), \(\eta\) (x), \(h_ i(x_ i)\) and weights \(d_ i>0 (i=1,...,p).\) All results reside on the representation \(R(\vartheta,\delta)=R(\vartheta,\delta^ 0)+2E_{\vartheta}[D(X)]\) for the risk function with a differential operator D(X); each solution of \(D(X)<0\) gives a dominating estimator \(\delta\). Special examples are given for the normal, gamma, and Dirichlet distributions.
      0 references
      construction of improved estimators
      0 references
      unified theory of simultaneous estimation
      0 references
      best invariant estimator
      0 references
      differential inequality
      0 references
      absolutely continuous
      0 references
      shrinkage estimators
      0 references
      arbitrary points
      0 references
      data based points
      0 references
      multivariate normal distribution
      0 references
      maximum likelihood
      0 references
      continuous exponential families
      0 references
      risk function
      0 references
      differential operator
      0 references
      dominating estimator
      0 references
      gamma
      0 references
      Dirichlet distributions
      0 references
      0 references

      Identifiers

      0 references
      0 references
      0 references
      0 references
      0 references
      0 references
      0 references