Projected principal component analysis in factor models
From MaRDI portal
Publication:5963521
Abstract: This paper introduces a Projected Principal Component Analysis (Projected-PCA), which employs principal component analysis to the projected (smoothed) data matrix onto a given linear space spanned by covariates. When it applies to high-dimensional factor analysis, the projection removes noise components. We show that the unobserved latent factors can be more accurately estimated than the conventional PCA if the projection is genuine, or more precisely, when the factor loading matrices are related to the projected linear space. When the dimensionality is large, the factors can be estimated accurately even when the sample size is finite. We propose a flexible semiparametric factor model, which decomposes the factor loading matrix into the component that can be explained by subject-specific covariates and the orthogonal residual component. The covariates' effects on the factor loadings are further modeled by the additive model via sieve approximations. By using the newly proposed Projected-PCA, the rates of convergence of the smooth factor loading matrices are obtained, which are much faster than those of the conventional factor analysis. The convergence is achieved even when the sample size is finite and is particularly appealing in the high-dimension-low-sample-size situation. This leads us to developing nonparametric tests on whether observed covariates have explaining powers on the loadings and whether they fully explain the loadings. The proposed method is illustrated by both simulated data and the returns of the components of the S&P 500 index.
Recommendations
- Robust projected principal component analysis for large-dimensional semiparametric factor modeling
- A self-reliant projected information criterion for the number of factors
- Using principal component analysis to estimate a high dimensional factor model with high-frequency data
- Directed principal component analysis
- Principal component analysis in very high-dimensional spaces
Cites work
- scientific article; zbMATH DE number 4048298 (Why is no real title available?)
- scientific article; zbMATH DE number 3765004 (Why is no real title available?)
- scientific article; zbMATH DE number 1911755 (Why is no real title available?)
- A Simple, Positive Semi-Definite, Heteroskedasticity and Autocorrelation Consistent Covariance Matrix
- A factor model approach to multiple testing under dependence
- A general framework for multiple testing dependence
- Asymptotics of sample eigenstructure for a large dimensional spiked covariance model
- Consistency of sparse PCA in high dimension, low sample size contexts
- Correlated \(z\)-values and the accuracy of large-scale statistical estimates
- Covariance regularization by thresholding
- Cross-dimensional inference of dependent high-dimensional data
- Determining the Number of Factors in Approximate Factor Models
- Determining the Number of Factors in the General Dynamic Factor Model
- Dynamic factor models with infinite-dimensional factor spaces: one-sided representations
- Efficient semiparametric estimation of the Fama-French model and extensions
- Eigenvalue ratio test for the number of factors
- Estimating false discovery proportion under arbitrary covariance dependence
- Estimation of (near) low-rank matrices with noise and high-dimensional scaling
- Exact matrix completion via convex optimization
- Factor modeling for high-dimensional time series: inference for the number of factors
- Forecasting Using Principal Components From a Large Number of Predictors
- GLS estimation of dynamic factor models
- Heteroskedasticity and Autocorrelation Consistent Covariance Matrix Estimation
- Improved penalization for determining the number of factors in approximate factor models
- Inferential Theory for Factor Models of Large Dimensions
- Nuclear-norm penalization and optimal rates for noisy low-rank matrix completion
- On the distribution of the largest eigenvalue in principal components analysis
- PCA consistency in high dimension, low sample size context
- Principal components estimation and identification of static factors
- Risks of large portfolios
- Sparse PCA: optimal rates and adaptive estimation
- Sparse principal component analysis and iterative thresholding
- Statistical analysis of factor models of high dimension
- Supervised singular value decomposition and its asymptotic properties
- The Generalized Dynamic Factor Model
- The high-dimension, low-sample-size geometric representation holds under mild conditions
- Time series modelling with semiparametric factor dynamics
Cited in
(46)- Interpretable Sparse Proximate Factors for Large Dimensions
- Sequential Scaled Sparse Factor Regression
- State-Varying Factor Models of Large Dimensions
- Dynamic Peer Groups of Arbitrage Characteristics
- Supervised singular value decomposition and its asymptotic properties
- Learning Latent Factors From Diversified Projections and Its Applications to Over-Estimated and Weak Factors
- Sparse online principal component analysis for parameter estimation in factor model
- Cross-Validated Loss-based Covariance Matrix Estimator Selection in High Dimensions
- Recent developments in high dimensional covariance estimation and its related issues, a review
- Rejoinder
- Continuum directions for supervised dimension reduction
- Large sample correlation matrices with unbounded spectrum
- Factor Extraction in Dynamic Factor Models: Kalman Filter Versus Principal Components
- Doubly debiased Lasso: high-dimensional inference under hidden confounding
- Uniform predictive inference for factor models with instrumental and idiosyncratic betas
- Time varying factor models with possibly strongly correlated noises
- Large covariance estimation through elliptical factor models
- Large dimensional latent factor modeling with missing observations and applications to causal inference
- Nonparametric fixed effects model for panel data with locally stationary regressors
- Changes in the span of systematic risk exposures
- Varying-Coefficient Panel Data Models With Nonstationarity and Partially Observed Factor Structure
- Estimating latent asset-pricing factors
- Non-asymptotic properties of spectral decomposition of large Gram-type matrices and applications
- Autoencoder asset pricing models
- Factor Models for High-Dimensional Tensor Time Series
- Augmented factor models with applications to validating market risk factors and forecasting bond risk premia
- Supervised multiway factorization
- High dimensional minimum variance portfolio estimation under statistical factor models
- Predicting the VIX and the volatility risk premium: the role of short-run funding spreads volatility factors
- Subfactor projections.
- Heterogeneity adjustment with applications to graphical model inference
- A semiparametric latent factor model for large scale temporal data with heteroscedasticity
- An integrated panel data approach to modelling economic growth
- A self-reliant projected information criterion for the number of factors
- Dynamic Semiparametric Factor Model With Structural Breaks
- Least squares estimation of large dimensional threshold factor models
- Quasi maximum likelihood analysis of high dimensional constrained factor models
- Robust covariance estimation for approximate factor models
- Semi-parametric single-index panel data models with interactive fixed effects: theory and practice
- Estimation and inference in semiparametric quantile factor models
- Directed principal component analysis
- Robust projected principal component analysis for large-dimensional semiparametric factor modeling
- Intraday cross-sectional distributions of systematic risk
- GMM estimation for high-dimensional panel data models
- Target PCA: transfer learning large dimensional panel data
- Sufficient forecasting using factor models
This page was built for publication: Projected principal component analysis in factor models
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5963521)