Greedy algorithms for prediction
From MaRDI portal
Publication:265302
Abstract: In many prediction problems, it is not uncommon that the number of variables used to construct a forecast is of the same order of magnitude as the sample size, if not larger. We then face the problem of constructing a prediction in the presence of potentially large estimation error. Control of the estimation error is either achieved by selecting variables or combining all the variables in some special way. This paper considers greedy algorithms to solve this problem. It is shown that the resulting estimators are consistent under weak conditions. In particular, the derived rates of convergence are either minimax or improve on the ones given in the literature allowing for dependence and unbounded regressors. Some versions of the algorithms provide fast solution to problems such as Lasso.
Recommendations
- 10.1162/153244304773936108
- Approximation and learning by greedy algorithms
- On the selection of predictors by using greedy algorithms and information theoretic criteria
- Coordinate descent algorithms for lasso penalized regression
- On the consistency of feature selection using greedy least squares regression
Cites Work
- scientific article; zbMATH DE number 845714 (Why is no real title available?)
- scientific article; zbMATH DE number 5056247 (Why is no real title available?)
- A Recursive Algorithm for Mixture of Densities Estimation
- A cross-validatory method for dependent data
- A dynamic model of expected bond returns: A functional gradient descent approach
- A maximal 𝕃_{𝕡}-inequality for stationary sequences and its applications
- A new covariance inequality and applications.
- A new mixing notion and functional central limit theorems for a sieve bootstrap in time series
- A new weak dependence condition and applications to moment inequalities
- A nonparametric estimator for the covariance function of functional data
- A simple lemma on greedy approximation in Hilbert space and convergence rates for projection pursuit regression and neural network training
- Aggregation for Gaussian regression
- An iterative thresholding algorithm for linear inverse problems with a sparsity constraint
- Approximation and learning by greedy algorithms
- Asymptotic theory of weakly dependent stochastic processes
- Basic properties of strong mixing conditions. A survey and some open questions
- Best subset selection, persistence in high-dimensional statistical learning and optimization under \(l_1\) constraint
- Boosting With theL2Loss
- Boosting for high-dimensional linear models
- Bootstrap model selection for possibly dependent and heterogeneous data
- Confidence sets in sparse regression
- Convergence of a block coordinate descent method for nondifferentiable minimization
- Coresets, sparse greedy approximation, and the Frank-Wolfe algorithm
- DATA-DEPENDENT ESTIMATION OF PREDICTION FUNCTIONS
- Degrees of freedom in lasso problems
- Estimating the Error Rate of a Prediction Rule: Improvement on Cross-Validation
- Forecast combination across estimation windows
- Forecasting Time Series Subject to Multiple Structural Breaks
- Gaussian model selection
- Greedy approximation
- High-dimensional generalized linear models and the lasso
- High-dimensional regression with noisy and missing data: provable guarantees with nonconvexity
- Introduction to nonparametric estimation
- Invariance principles for absolutely regular empirical processes
- Learning Theory and Kernel Machines
- Least angle regression. (With discussion)
- Matching pursuits with time-frequency dictionaries
- Maximal inequalities via bracketing with adaptive truncation
- Minimax estimation via wavelet shrinkage
- Mixing properties of ARMA processes
- Nearly unbiased variable selection under minimax concave penalty
- Non-strong mixing autoregressive processes
- On Measuring and Correcting the Effects of Data Mining and Model Selection
- On asymptotically optimal confidence regions and tests for high-dimensional models
- On the ``degrees of freedom of the lasso
- On the consistency of feature selection using greedy least squares regression
- On the uniform convergence of empirical norms and inner products, with application to causal inference
- Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise
- Pathwise coordinate optimization
- Persistene in high-dimensional linear predictor-selection and the virtue of overparametrization
- Propriétés de mélange des processus autorégressifs polynomiaux. (Mixing properties of polynomial autoregressive processes)
- Rate of convergence of pure greedy algorithm.
- Regular variation of GARCH processes.
- Risk bounds for mixture density estimation
- Sieve Extremum Estimates for Weakly Dependent Data
- Smoothing Parameter Selection in Nonparametric Regression Using an Improved Akaike Information Criterion
- Some remarks on greedy algorithms
- Sparse models and methods for optimal instruments with an application to eminent domain
- Sparsity oracle inequalities for the Lasso
- Splines for financial volatility
- Statistical significance in high-dimensional linear models
- Statistics for high-dimensional data. Methods, theory and applications.
- The Dantzig selector: statistical estimation when \(p\) is much larger than \(n\). (With discussions and rejoinder).
- Two lower estimates in greedy approximation
- Universal approximation bounds for superpositions of a sigmoidal function
- Weak greedy algorithms
- \(\ell _{1}\)-regularized linear regression: persistence and oracle inequalities
- \(\ell_1\)-penalized quantile regression in high-dimensional sparse models
Cited In (10)
- Orthogonal one step greedy procedure for heteroscedastic linear models
- Model selection for high-dimensional linear regression with dependent observations
- Efficient prediction algorithms for binary decomposition techniques
- 10.1162/153244304773936108
- Prediction-based termination rule for greedy learning with massive data
- On the selection of predictors by using greedy algorithms and information theoretic criteria
- Semiparametric estimation of plane similarities: application to fast computation of aeronautic loads
- Algorithms with Predictions
- On the consistency of feature selection using greedy least squares regression
- Estimation for the prediction of point processes with many covariates
This page was built for publication: Greedy algorithms for prediction
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q265302)