Estimating the Error Rate of a Prediction Rule: Improvement on Cross-Validation

From MaRDI portal
Publication:3332119

DOI10.2307/2288636zbMath0543.62079OpenAlexW4243562335MaRDI QIDQ3332119

Efron, Bradley

Publication date: 1983

Full work available at URL: https://doi.org/10.2307/2288636



Related Items

Greedy algorithms for prediction, A new variable selection approach using random forests, Post-hocanalyses in multiple regression based on prediction error, Combining neural networks for function approximation under conditions of sparse data: the biased regression approach, Integration of rough set and neural network ensemble to predict the configuration performance of a modular product family, A shrunken leaving-one-out estimator of error rate, On stability issues in deriving multivariable regression models, Large sample properties of cross-validation assessment statistics, A bootstrap generalized likelihood ratio test in discriminant analysis, Evaluating the impact of exploratory procedures in regression prediction: A pseudosample approach, Model selection criteria based on cross-validatory concordance statistics, Extending AIC to best subset regression, Iterative Bias Correction of the Cross-Validation Criterion, A serial risk score approach to disease classification that accounts for accuracy and cost, On the biases of error estimators in prediction problems, Model-free model-fitting and predictive distributions, The use of Smooth Bootstrap Techniques for Estimating the Error Rate of a Prediction Rule, Surrogates for numerical simulations; optimization of eddy-promoter heat exchangers, Assessing the performance of an allocation rule, The robust estimation of classification error rates, Bootstrap by sequential resampling, Resampling-based information criteria for best-subset regression, Variable selection for generalized linear mixed models by \(L_1\)-penalized estimation, Model building using bi-level optimization, Optimal mean-square-error calibration of classifier error estimators under Bayesian models, Bayesian multiscale smoothing in supervised and semi-supervised kernel discriminant analysis, A note on estimating the msep in nonlinear regression, Model selection for two-sample problems with right-censored data: an application of Cox model, Bias and variance reduction techniques for bootstrap information criteria, Bootstrap model selection for possibly dependent and heterogeneous data, Bootstrap-based model selection criteria for beta regressions, Order-based error for managing ensembles of surrogates in mesh adaptive direct search, Optimal model selection in heteroscedastic regression using piecewise polynomial functions, Model selection by resampling penalization, Sparse supervised dimension reduction in high dimensional classification, The negative correlations between data-determined bandwidths and the optimal bandwidth, Tree-structured prognostic classification for censored survival data: validation of computationally inexpensive model selection criteria, Margin-adaptive model selection in statistical learning, Estimation of error rate for linear discriminant functions by resampling: Non-Gaussian populations, Resampling-type error rate estimation for linear discriminant functions. Pearson VII distributions, Measuring the prediction error. A comparison of cross-validation, bootstrap and covariance penalty methods, Fast robust estimation of prediction error based on resampling, Asymptotic optimality of full cross-validation for selecting linear regression models, A large-sample model selection criterion based on Kullback's symmetric divergence, Comment: Automated analyses: because we can, does it mean we should?, Multiple group linear discriminant analysis: robustness and error rate, Provably safe and robust learning-based model predictive control, Unsupervised stratification of cross-validation for accuracy estimation, Cross-Validation, Risk Estimation, and Model Selection: Comment on a Paper by Rosset and Tibshirani, An assumption for the development of bootstrap variants of the Akaike information criterion in mixed models, Optimal Combinations of Diagnostic Tests Based on AUC, A comparison of parametric conditional error-rate estimators for the two-group linear discriminant function, Estimating recorder points and other management science applications by bootstrap procedure, PLS generalised linear regression, Towards optimal use of incomplete classification data, Degrees of freedom in low rank matrix estimation, Bootstrap estimated true and false positive rates and ROC curve, Multivariate exponential survival trees and their application to tooth prognosis, Estimating classification error rate: repeated cross-validation, repeated hold-out and bootstrap, Conditional confidence intervals for classification error rate, Model combination for credit risk assessment: a stacked generalization approach, Estimating Prediction Error: Cross-Validation vs. Accumulated Prediction Error, Bolstered error estimation, Multi-class tumor classification by discriminant partial least squares using microarray gene expression data and assessment of classification models, A survey of cross-validation procedures for model selection, A gamma kernel density estimation for insurance loss data, Maximizing proportions of correct classifications in binary logistic regression, Asymptotics of cross-validated risk estimation in estimator selection and performance assess\-ment, Cross validation model selection criteria for linear regression based on the Kullback-Leibler discrepancy, Selection bias in working with the top genes in supervised classification of tissue samples, Optimal convex error estimators for classification, Bandwidth choice for nonparametric classification, Ideal point discriminant analysis, Improved model identification for non-linear systems using a random subsampling and multifold modelling (RSMM) approach, An improved kernel regression method based on Taylor expansion, Locally weighted regression models for surrogate-assisted design optimization, Bayesian nonparametric model selection and model testing, Model selection by normalized maximum likelihood, Empirical comparison of error rate-estimators in logistic discriminant analysis, New Bootstrap Applications in Supervised Learning, A multistage algorithm for best-subset model selection based on the Kullback-Leibler discrepancy, Model evaluation, discrepancy function estimation, and social choice theory, Assessing classifiers in terms of the partial area under the ROC curve, ROS regression: integrating regularization with optimal scaling regression, Assessing incremental value of biomarkers with multi-phase nested case-control studies, Weakly supervised clustering: learning fine-grained signals from coarse labels, Fast and robust bootstrap, Bootstrap variants of the Akaike information criterion for mixed model selection, Selection of regressors in econometrics: parametric and nonparametric methods selection of regressors in econometrics, Validation in principal components analysis applied to EEG data, On the predictive risk in misspecified quantile regression, Exact performance of error estimators for discrete classifiers, Asymptotic bootstrap corrections of AIC for linear regression models, On model selection via stochastic complexity in robust linear regression, Estimating prediction error in microarray classification: Modifications of the 0.632+ bootstrap when ${\bf n} < {\bf p}$, A piecewise transition model for analyzing multistate life history data, The use of a distance measure in regularised discriminant analysis, Comparing and selecting spatial predictors using local criteria, Cross-conformal predictors, Re-sampling procedures for reducing bias of error rate estimation in multinomial classification, A flexible and efficient spatial interpolator for radar rainfall estimation, Data-based interval estimation of classification error rates, The comparison study of the model selection criteria on the Tobit regression model based on the bootstrap sample augmentation mechanisms, Comparison of classical tumour growth models for patient derived and cell-line derived xenografts using the nonlinear mixed-effects framework, Model selection in linear regression using paired bootstrap, Ten More Years of Error Rate Research, Bootstrap estimation and model selection for multivariate normal mixtures using parallel computing with graphics processing units, On the estimation of prediction errors in logistic regression models, Discriminant analysis with stratified prior probabilities, Estimating the Kullback–Liebler risk based on multifold cross‐validation, Clinical risk prediction models and informative cluster size: Assessing the performance of a suicide risk prediction algorithm, Three distributions in the extended occupancy problem, The fraud loss for selecting the model complexity in fraud detection, Criterion constrained Bayesian hierarchical models, Assessing the variability of posterior probabilities in Gaussian model-based clustering, Estimationg error rates in discriminant analysis with correlated training observations: a simulation study, Autoregressive model selection based on a prediction perspective, Using Multilayer Perceptron Computation to Discover Ideal Insect Olfactory Receptor Combinations in the Mosquito and Fruit Fly for an Efficient Electronic Nose, Sequential regression measurement error models with application, Comparison of procedures to assess non‐linear and time‐varying effects in multivariable models for survival data, Determination of the best significance level in forward stepwise logistic regression, The asymptotic distribution of the proportion of correct classifications for a holdout sample in logistic regression, Evaluating Decision Rules for Nitrogen Fertilization, A regression model selection criterion based on bootstrap bumping for use with resistant fitting., Feature selection algorithms in classification problems: an experimental evaluation, Prediction risk for the horseshoe regression, Efron‐Type Measures of Prediction Error for Survival Analysis, Determination of the Selection Statistics and Best Significance Level in Backward Stepwise Logistic Regression, Statistical significance of the Netflix challenge, Recent developments in bootstrap methodology, Cross‐validation and peeling strategies for survival bump hunting using recursive peeling methods, Adaptive order selection for autoregressive models, Determination of different types of fixed effects in three-dimensional panels*, Reconceptualizing the p -value from a likelihood ratio test: a probabilistic pairwise comparison of models based on Kullback-Leibler discrepancy measures, A Modified Akaike Criterion for Model Choice in Generalized Linear Models, Statistical Inference Enables Bad Science; Statistical Thinking Enables Good Science