Limiting values of large deviation probabilities of quadratic statistics (Q750050)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Limiting values of large deviation probabilities of quadratic statistics |
scientific article |
Statements
Limiting values of large deviation probabilities of quadratic statistics (English)
0 references
1990
0 references
If T(P) is a functional and \(T_ n=T(\hat P_ n)\) is used as a test statistic then the application of Bahadur efficiency leads to the expression \[ K(\Omega_{\epsilon},P)=\inf \{K(Q,P):\;T(Q)-T(P)>\epsilon \}, \] where K is the Kullback-Leibler information number. But \(K(\Omega_{\epsilon},P)\) is often complicate. Therefore the authors study the behaviour of \(K(\Omega_{\epsilon},P)/\epsilon\) as \(\epsilon\downarrow 0\) for quadratic functionals \(\iint \psi (s,t)dQ(s)dQ(t).\) One typical result is Theorem 3.3, which asserts that \[ \lim_{\epsilon \downarrow 0}K(\Omega_{\epsilon},P)/\epsilon =(2\lambda_ 1)^{-1}, \] where \(\lambda_ 1\) is the largest eigenvalue of the integral operator defined by the kernel \(\psi\). Applications of the results concern the generalized Cramér-von Mises statistic, Neyman's smooth test and likelihood ratio tests.
0 references
local limits of large deviation probabilities
0 references
sums of k-dimensional random vectors
0 references
Anderson-Darling statistic
0 references
quadratic statistics
0 references
Hilbert-Schmidt operator
0 references
eigenfunctions
0 references
likelihood ratio tests
0 references
Bahadur efficiency
0 references
Kullback-Leibler information
0 references
quadratic functionals
0 references
largest eigenvalue
0 references
integral operator
0 references
generalized Cramér-von Mises statistic
0 references
Neyman's smooth test
0 references
0 references
0 references
0 references