Sparse recovery under weak moment assumptions
From MaRDI portal
Publication:520739
DOI10.4171/JEMS/682zbMath1414.62135arXiv1401.2188WikidataQ105584474 ScholiaQ105584474MaRDI QIDQ520739
Shahar Mendelson, Guillaume Lecué
Publication date: 5 April 2017
Published in: Journal of the European Mathematical Society (JEMS) (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1401.2188
Related Items
On Multiplier Processes Under Weak Moment Assumptions, The gap between the null space property and the restricted isometry property, Performance of empirical risk minimization in linear aggregation, Controlling the least eigenvalue of a random Gram matrix, On the interval of fluctuation of the singular values of random matrices, Improved bounds for sparse recovery from subsampled random convolutions, Conjugate gradient acceleration of iteratively re-weighted least squares methods, On the geometry of polytopes generated by heavy-tailed random vectors, Low rank matrix recovery from rank one measurements, Generic error bounds for the generalized Lasso with sub-exponential data, Sample average approximation with heavier tails II: localization in stochastic convex optimization and persistence results for the Lasso, Estimation of the \(\ell_2\)-norm and testing in sparse linear regression with unknown variance, Regularization, sparse recovery, and median-of-means tournaments, A Rice method proof of the null-space property over the Grassmannian, Flavors of Compressive Sensing, Column normalization of a random measurement matrix, Slope meets Lasso: improved oracle bounds and optimality, Dimensionality reduction with subgaussian matrices: a unified theory, Regularization and the small-ball method. I: Sparse recovery, Learning from MOM's principles: Le Cam's approach, Learning without Concentration, Phase retrieval with PhaseLift algorithm, Estimation in High Dimensions: A Geometric Perspective, Non-Gaussian hyperplane tessellations and robust one-bit compressed sensing, Maximin effects in inhomogeneous large-scale data, Sparse recovery from extreme eigenvalues deviation inequalities, Convergence rates of least squares regression estimators with heavy-tailed errors, Regularization and the small-ball method II: complexity dependent error rates, Preserving injectivity under subgaussian mappings and its application to compressed sensing
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- The lower tail of random quadratic forms with applications to ordinary least squares
- A mathematical introduction to compressive sensing
- Covariance estimation for distributions with \({2+\varepsilon}\) moments
- On higher order isotropy conditions and lower bounds for sparse quadratic forms
- Statistics for high-dimensional data. Methods, theory and applications.
- Oracle inequalities in empirical risk minimization and sparse recovery problems. École d'Été de Probabilités de Saint-Flour XXXVIII-2008.
- Restricted isometry property of matrices with independent columns and neighborly polytopes by random sampling
- Concentration inequalities and model selection. Ecole d'Eté de Probabilités de Saint-Flour XXXIII -- 2003.
- The restricted isometry property and its implications for compressed sensing
- Uniform uncertainty principle for Bernoulli and subgaussian ensembles
- Central limit theorems for empirical measures
- Estimation of moments of sums of independent real random variables
- Weak convergence and empirical processes. With applications to statistics
- Simultaneous analysis of Lasso and Dantzig selector
- Stability and robustness of \(\ell_1\)-minimizations with Weibull matrices and redundant dictionaries
- The Dantzig selector: statistical estimation when \(p\) is much larger than \(n\). (With discussions and rejoinder).
- Atomic Decomposition by Basis Pursuit
- Learning without Concentration
- Reconstruction From Anisotropic Random Measurements
- Small Ball Probabilities for Linear Images of High-Dimensional Distributions
- Bounding the Smallest Singular Value of a Random Matrix Without Concentration
- On sparse reconstruction from Fourier and Gaussian measurements
- Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information
- Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?
- Cube Slicing in R n
- Linear Inversion of Band-Limited Reflection Seismograms
- Signal Recovery and the Large Sieve
- Uncertainty principles and ideal atomic decomposition
- On tight bounds for the Lasso
- Sparse Approximate Solutions to Linear Systems
- A Remark on the Diameter of Random Sections of Convex Bodies
- Weakly decomposable regularization penalties and structured sparsity
- Optimally sparse representation in general (nonorthogonal) dictionaries via ℓ 1 minimization
- Stable signal recovery from incomplete and inaccurate measurements
- Compressed sensing
- Gaussian model selection