On two simple and effective procedures for high dimensional classification of general populations
From MaRDI portal
Publication:284189
DOI10.1007/S00362-015-0660-8zbMATH Open1336.62187arXiv1501.01763OpenAlexW2076636512MaRDI QIDQ284189FDOQ284189
Authors: Zhaoyuan Li, Jian-Feng Yao
Publication date: 17 May 2016
Published in: Statistical Papers (Search for Journal in Brave)
Abstract: In this paper, we generalize two criteria, the determinant-based and trace-based criteria proposed by Saranadasa (1993), to general populations for high dimensional classification. These two criteria compare some distances between a new observation and several different known groups. The determinant-based criterion performs well for correlated variables by integrating the covariance structure and is competitive to many other existing rules. The criterion however requires the measurement dimension be smaller than the sample size. The trace-based criterion in contrast, is an independence rule and effective in the "large dimension-small sample size" scenario. An appealing property of these two criteria is that their implementation is straightforward and there is no need for preliminary variable selection or use of turning parameters. Their asymptotic misclassification probabilities are derived using the theory of large dimensional random matrices. Their competitive performances are illustrated by intensive Monte Carlo experiments and a real data analysis.
Full work available at URL: https://arxiv.org/abs/1501.01763
Recommendations
- Non-parametric comparison and classification of two large-scale populations
- High breakdown estimation for multiple populations with applications to discriminant analysis
- Classifying two populations by Bayesian method and applications
- Neyman-Pearson classification under high-dimensional settings
- Stein's method in high dimensional classification and applications
- Characterizing the scale dimension of a high-dimensional classification problem
- Asymptotic expansion of the misclassification probabilities of D- and A- criteria for discrimination from two high dimensional populations using the theory of large dimensional random matrices
- Asymptotic probabilities of misclassification of two discriminant functions in cases of high dimensional data
- Classification and Generalized Principal Component Analysis
delocalizationdeterminant-based criterionhigh dimensional classificationlarge sample covariance matrixtrace-based criterion
Cites Work
- Regularized linear discriminant analysis and its application in microarrays
- Fast nonparametric classification based on data depth
- Title not available (Why is that?)
- High-dimensional classification using features annealed independence rules
- Spectral analysis of large dimensional random matrices
- Two sample tests for high-dimensional covariance matrices
- Some theory for Fisher's linear discriminant function, `naive Bayes', and some alternatives when there are many more variables than observations
- Sparse linear discriminant analysis by thresholding for high dimensional data
- Title not available (Why is that?)
- Tests for high-dimensional covariance matrices
- A Road to Classification in High Dimensional Space: The Regularized Optimal Affine Discriminant
- Asymptotic expansion of the misclassification probabilities of D- and A- criteria for discrimination from two high dimensional populations using the theory of large dimensional random matrices
- Asymptotic probabilities of misclassification of two discriminant functions in cases of high dimensional data
- Enhancement of the applicability of Markowitz's portfolio optimization by utilizing random matrix theory
- Discriminant analysis of multivariate repeated measures data with Kronecker product structured covariance matrices
- Some tests for the covariance matrix with fewer observations than the dimension under non-normality
- Error rates in classification consisting of discrete and continuous variables in the presence of covariates
Cited In (5)
- Two-group classification with high-dimensional correlated data: a factor model approach
- Test on the linear combinations of covariance matrices in high-dimensional data
- On the dimension effect of regularized linear discriminant analysis
- High-dimensional linear models: a random matrix perspective
- Asymptotic expansion of the misclassification probabilities of D- and A- criteria for discrimination from two high dimensional populations using the theory of large dimensional random matrices
Uses Software
This page was built for publication: On two simple and effective procedures for high dimensional classification of general populations
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q284189)