On a Class of Optimization-Based Robust Estimators
From MaRDI portal
Publication:4566947
Abstract: We consider in this paper the problem of estimating a parameter matrix from observations which are affected by two types of noise components: (i) a sparse noise sequence which, whenever nonzero can have arbitrarily large amplitude (ii) and a dense and bounded noise sequence of "moderate" amount. This is termed a robust regression problem. To tackle it, a quite general optimization-based framework is proposed and analyzed. When only the sparse noise is present, a sufficient bound is derived on the number of nonzero elements in the sparse noise sequence that can be accommodated by the estimator while still returning the true parameter matrix. While almost all the restricted isometry-based bounds from the literature are not verifiable, our bound can be easily computed through solving a convex optimization problem. Moreover, empirical evidence tends to suggest that it is generally tight. If in addition to the sparse noise sequence, the training data are affected by a bounded dense noise, we derive an upper bound on the estimation error.
Cited in
(9)- SOCP based variance free Dantzig selector with application to robust estimation
- A stochastic analysis of robust estimation algorithms inH∞with rational basis functions
- A Competitive Minimax Approach to Robust Estimation of Random Parameters
- On sparsity‐inducing methods in system identification and state estimation
- Robust estimation via generalized quasi-gradients
- Robust estimator design with an emphasis balance between performance and robustness
- \(\sqrt n\)-consistent robust integration-based estimation
- scientific article; zbMATH DE number 6785488 (Why is no real title available?)
- Robustness analysis of a maximum correntropy framework for linear regression
This page was built for publication: On a Class of Optimization-Based Robust Estimators
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4566947)