Subspace clustering of high-dimensional data: a predictive approach
From MaRDI portal
(Redirected from Publication:740601)
Abstract: In several application domains, high-dimensional observations are collected and then analysed in search for naturally occurring data clusters which might provide further insights about the nature of the problem. In this paper we describe a new approach for partitioning such high-dimensional data. Our assumption is that, within each cluster, the data can be approximated well by a linear subspace estimated by means of a principal component analysis (PCA). The proposed algorithm, Predictive Subspace Clustering (PSC) partitions the data into clusters while simultaneously estimating cluster-wise PCA parameters. The algorithm minimises an objective function that depends upon a new measure of influence for PCA models. A penalised version of the algorithm is also described for carrying our simultaneous subspace clustering and variable selection. The convergence of PSC is discussed in detail, and extensive simulation results and comparisons to competing methods are presented. The comparative performance of PSC has been assessed on six real gene expression data sets for which PSC often provides state-of-art results.
Recommendations
Cites work
- scientific article; zbMATH DE number 3984329 (Why is no real title available?)
- scientific article; zbMATH DE number 3753890 (Why is no real title available?)
- scientific article; zbMATH DE number 845714 (Why is no real title available?)
- A framework for feature selection in clustering
- A survey on enhanced subspace clustering
- Adjustment of an Inverse Matrix Corresponding to a Change in One Element of a Given Matrix
- Asymptotic convergence analysis of the projection approximation subspace tracking algorithms
- Consensus clustering: A resampling-based method for class discovery and visualization of gene expression microarray data
- Estimating the number of clusters in a data set via the gap statistic
- Hybrid linear modeling via local best-fit flats
- Identifying multiple influential observations in linear regression
- Influential observations, high leverage points, and outliers in linear regression
- Pathwise coordinate optimization
- Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$-Constrained Quadratic Programming (Lasso)
- Sparse principal component analysis via regularized low rank matrix approximation
- Subspace clustering of high-dimensional data: a predictive approach
- \(k\)-plane clustering
Cited in
(16)- Partition clustering of high dimensional low sample size data based on \(p\)-values
- Robust subspace clustering
- Analyzing gene expression data with predictive clustering trees
- Is-ClusterMPP: clustering algorithm through point processes and influence space towards high-dimensional data
- Dense community detection in multi-valued attributed networks
- Projective clustering based on Parzen window technique
- Lossy compression approach to subspace clustering
- Subspace clustering of high-dimensional data: a predictive approach
- Subspace clustering for panel data with interactive effects
- Weighted sparse simplex representation: a unified framework for subspace clustering, constrained clustering, and active learning
- High dimensional data clustering from a dynamical systems point of view
- Hierarchical subspace clustering
- Subspace clustering with automatic feature grouping
- Enhancing principal direction divisive clustering
- Influential features PCA for high dimensional clustering
- The generic subspace clustering model
This page was built for publication: Subspace clustering of high-dimensional data: a predictive approach
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q740601)