Rates of convergence of minimum distance estimators and Kolmogorov's entropy
From MaRDI portal
Publication:1064711
DOI10.1214/aos/1176349553zbMath0576.62057MaRDI QIDQ1064711
Publication date: 1985
Published in: The Annals of Statistics (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1214/aos/1176349553
rate of convergence; density estimation; Kolmogorov entropy; empirical measure; Robust minimum distance estimators
Related Items
A general lower bound of minimax risk for absolute‐error loss, A note on penalized minimum distance estimation in nonparametric regression, Estimation and selection procedures in regression: anL1approach, Almost sure classification of densities, Parameter selection in modified histogram estimates, Density estimation by the penalized combinatorial method, Strongly consistent model selection for densities, \(L_ 1\)-optimal estimates for a regression type function in \(R^ d\), A universally acceptable smoothing factor for kernel density estimates, Universal smoothing factor selection in density estimation: theory and practice. (With discussion), Rates of convergence of estimates, Kolmogorov's entropy and the dimensionality reduction principle in regression, Nonasymptotic universal smoothing factors, kernel complexity and Yatracos classes, Information-theoretic determination of minimax rates of convergence, Dependence and the dimensionality reduction principle, Minimum distance regression-type estimates with rates under weak dependence, A theory of transfer learning with applications to active learning, A note on minimum distance estimation of copula densities, Optimal \(L_{1}\) bandwidth selection for variable kernel density estimates, LOCALIZED MODEL SELECTION FOR REGRESSION, A note on L1consistent estimation, Nonparametric density estimates with improved . performance on given sets of densities