Recovery guarantees for polynomial coefficients from weakly dependent data with outliers
From MaRDI portal
Publication:2209293
DOI10.1016/J.JAT.2020.105472zbMATH Open1451.41013OpenAlexW3071135787MaRDI QIDQ2209293FDOQ2209293
Authors: Lam Si Tung Ho, Hayden Schaeffer, Giang Tran, Rachel Ward
Publication date: 30 October 2020
Published in: Journal of Approximation Theory (Search for Journal in Brave)
Abstract: Learning non-linear systems from noisy, limited, and/or dependent data is an important task across various scientific fields including statistics, engineering, computer science, mathematics, and many more. In general, this learning task is ill-posed; however, additional information about the data's structure or on the behavior of the unknown function can make the task well-posed. In this work, we study the problem of learning nonlinear functions from corrupted and dependent data. The learning problem is recast as a sparse robust linear regression problem where we incorporate both the unknown coefficients and the corruptions in a basis pursuit framework. The main contribution of our paper is to provide a reconstruction guarantee for the associated -optimization problem where the sampling matrix is formed from dependent data. Specifically, we prove that the sampling matrix satisfies the null space property and the stable null space property, provided that the data is compact and satisfies a suitable concentration inequality. We show that our recovery results are applicable to various types of dependent data such as exponentially strongly -mixing data, geometrically -mixing data, and uniformly ergodic Markov chain. Our theoretical results are verified via several numerical simulations.
Full work available at URL: https://arxiv.org/abs/1811.10115
Recommendations
- SPARSE AND ROBUST LINEAR REGRESSION: AN OPTIMIZATION ALGORITHM AND ITS STATISTICAL PROPERTIES
- Learning sparse polynomial functions
- Correcting data corruption errors for multivariate function approximation
- Signal recovery from incomplete measurements in the presence of outliers
- Learning non-parametric basis independent models from point queries via low-rank methods
Multidimensional problems (41A63) Approximation by polynomials (41A10) Approximations and expansions (41A99)
Cites Work
- The elements of statistical learning. Data mining, inference, and prediction
- Regularization algorithms for learning that are equivalent to multilayer networks
- Splitting Algorithms for the Sum of Two Nonlinear Operators
- Nonlinear Regression with Dependent Observations
- Empirical Processes with Applications to Statistics
- Title not available (Why is that?)
- Deep learning
- Discovering governing equations from data by sparse identification of nonlinear dynamical systems
- Machine learning: Trends, perspectives, and prospects
- A CENTRAL LIMIT THEOREM AND A STRONG MIXING CONDITION
- Title not available (Why is that?)
- Proximal splitting methods in signal processing
- Title not available (Why is that?)
- A mathematical introduction to compressive sensing
- System identification in the presence of outliers and random noises: a compressed sensing approach
- High-dimensional adaptive sparse polynomial interpolation and applications to parametric PDEs
- Learning functions of few arbitrary linear parameters in high dimensions
- Compressed sensing and matrix completion with constant proportion of corruptions
- Effectively Well-Conditioned Linear Systems
- Title not available (Why is that?)
- Pattern recognition for conditionally independent data
- Interpolation via weighted \(\ell_{1}\) minimization
- A Bernstein-type inequality for some mixing processes and dynamical systems with an application to learning
- Approximation of infinitely differentiable multivariate functions is intractable
- Corrupted Sensing: Novel Guarantees for Separating Structured Signals
- Minimum complexity regression estimation with weakly dependent observations
- Learning from dependent observations
- EXPONENTIAL INEQUALITIES AND FUNCTIONAL ESTIMATIONS FOR WEAK DEPENDENT DATA: APPLICATIONS TO DYNAMICAL SYSTEMS
- On the convergence of the SINDy algorithm
- Modeling and nonlinear parameter estimation with Kronecker product representation for coupled oscillators and spatiotemporal systems
- Uniform Recovery Bounds for Structured Random Matrices in Corrupted Compressed Sensing
- Hidden physics models: machine learning of nonlinear partial differential equations
- Optimal weighted least-squares methods
- Polynomial approximation via compressed sensing of high-dimensional functions on lower sets
- Extracting Sparse High-Dimensional Dynamics from Limited Data
- Learning partial differential equations via data discovery and sparse optimization
- Sparse identification of nonlinear dynamics for model predictive control in the low-data limit
- Fast learning from \(\alpha\)-mixing observations
- Learning from non-iid data: fast rates for the one-vs-all multiclass plug-in classifiers
- Sequential function approximation with noisy data
- Generalization and Robustness of Batched Weighted Average Algorithm with V-Geometrically Ergodic Markov Data
- Compressed sensing with sparse corruptions: fault-tolerant sparse collocation approximations
Cited In (7)
- Correcting data corruption errors for multivariate function approximation
- Towards optimal sampling for learning sparse approximation in high dimensions
- Adaptive group Lasso neural network models for functions of few variables and time-dependent data
- Extracting Structured Dynamical Systems Using Sparse Optimization With Very Few Samples
- Learning from non-random data in Hilbert spaces: an optimal recovery perspective
- A generalization bound of deep neural networks for dependent data
- Model selection of chaotic systems from data with hidden variables using sparse data assimilation
Uses Software
This page was built for publication: Recovery guarantees for polynomial coefficients from weakly dependent data with outliers
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2209293)