Impacts of high dimensionality in finite samples
From MaRDI portal
Abstract: High-dimensional data sets are commonly collected in many contemporary applications arising in various fields of scientific research. We present two views of finite samples in high dimensions: a probabilistic one and a nonprobabilistic one. With the probabilistic view, we establish the concentration property and robust spark bound for large random design matrix generated from elliptical distributions, with the former related to the sure screening property and the latter related to sparse model identifiability. An interesting concentration phenomenon in high dimensions is revealed. With the nonprobabilistic view, we derive general bounds on dimensionality with some distance constraint on sparse models. These results provide new insights into the impacts of high dimensionality in finite samples.
Recommendations
Cites work
- scientific article; zbMATH DE number 3886886 (Why is no real title available?)
- scientific article; zbMATH DE number 45785 (Why is no real title available?)
- scientific article; zbMATH DE number 2174437 (Why is no real title available?)
- A Selective Overview of Variable Selection in High Dimensional Feature Space (Invited Review Article)
- A characterization of the distributions that imply mean-variance utility functions
- A unified approach to model selection and sparse recovery using regularized least squares
- DIFFERENTIAL GEOMETRY OF GRASSMANN MANIFOLDS
- Effect of heavy tails on ultra high dimensional variable ranking methods
- Feature screening via distance correlation learning
- Geometric Representation of High Dimension, Low Sample Size Data
- High-dimensional classification using features annealed independence rules
- High-dimensional variable screening and bias in subsequent inference, with an empirical comparison
- Least angle regression. (With discussion)
- Model-free feature screening for ultrahigh-dimensional data
- Nonparametric independence screening in sparse ultra-high-dimensional additive models
- Normal Multivariate Analysis and the Orthogonal Group
- On the conditions used to prove oracle results for the Lasso
- One-step sparse estimates in nonconcave penalized likelihood models
- Optimally sparse representation in general (nonorthogonal) dictionaries via ℓ 1 minimization
- Packing Lines, Planes, etc.: Packings in Grassmannian Spaces
- Simultaneous analysis of Lasso and Dantzig selector
- Statistical challenges with high dimensionality: feature selection in knowledge discovery
- Sure independence screening and compressed random sensing
- The Geometry of Algorithms with Orthogonality Constraints
- The Kolmogorov filter for variable screening in high-dimensional binary classification
- The concentration of measure phenomenon
- Tilting methods for assessing the influence of components in a classifier
Cited in
(7)- A survey of high dimension low sample size asymptotics
- IPAD: stable interpretable forecasting with knockoffs inference
- RANK: Large-Scale Inference With Graphical Nonlinear Knockoffs
- Nonsparse learning with latent variables
- Greedy forward regression for variable screening
- Statistical insights into deep neural network learning in subspace classification
- A fundamental bias in calculating dimensions from finite data sets
This page was built for publication: Impacts of high dimensionality in finite samples
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q385798)