Matrix Factor Analysis: From Least Squares to Iterative Projection
From MaRDI portal
Publication:6150367
DOI10.1080/07350015.2023.2191676arXiv2112.04186OpenAlexW4327587455MaRDI QIDQ6150367FDOQ6150367
Authors: Yong He, Xin-Bing Kong, Long Yu, Xinsheng Zhang, Changwei Zhao
Publication date: 6 March 2024
Published in: Journal of Business and Economic Statistics (Search for Journal in Brave)
Abstract: In this article, we study large-dimensional matrix factor models and estimate the factor loading matrices and factor score matrix by minimizing square loss function. Interestingly, the resultant estimators coincide with the Projected Estimators (PE) in Yu et al.(2022), which was proposed from the perspective of simultaneous reduction of the dimensionality and the magnitudes of the idiosyncratic error matrix. In other word, we provide a least-square interpretation of the PE for matrix factor model, which parallels to the least-square interpretation of the PCA for the vector factor model. We derive the convergence rates of the theoretical minimizers under sub-Gaussian tails. Considering the robustness to the heavy tails of the idiosyncratic errors, we extend the least squares to minimizing the Huber loss function, which leads to a weighted iterative projection approach to compute and learn the parameters. We also derive the convergence rates of the theoretical minimizers of the Huber loss function under bounded th moment of the idiosyncratic errors. We conduct extensive numerical studies to investigate the empirical performance of the proposed Huber estimators relative to the state-of-the-art ones. The Huber estimators perform robustly and much better than existing ones when the data are heavy-tailed, and as a result can be used as a safe replacement in practice. An application to a Fama-French financial portfolio dataset demonstrates the empirical advantage of the Huber estimator.
Full work available at URL: https://arxiv.org/abs/2112.04186
Cites Work
- Forecasting Using Principal Components From a Large Number of Predictors
- Inferential Theory for Factor Models of Large Dimensions
- Determining the Number of Factors in Approximate Factor Models
- Principal components estimation and identification of static factors
- Eigenvalue ratio test for the number of factors
- Robust Estimation of a Location Parameter
- Large Covariance Estimation by Thresholding Principal Orthogonal Complements
- Projected estimation for large-dimensional matrix factor models
- Factor models for matrix-valued high-dimensional time series
- Rank determination in tensor factor model
- High-dimensional statistics. A non-asymptotic viewpoint
- Identification and estimation of threshold matrix-variate factor models
- Factor Models for High-Dimensional Tensor Time Series
- Testing hypotheses about the number of factors in large factor models
- Quantile factor models
- Separable factor analysis with applications to mortality data
- Constrained Factor Models for High-Dimensional Matrix-Variate Time Series
- Towards a universal self-normalized moderate deviation
- Self-normalized large deviations
- Robust factor number specification for large-dimensional elliptical factor model
- Using principal component analysis to estimate a high dimensional factor model with high-frequency data
- High-frequency factor models and regressions
- Community detection on mixture multilayer networks via regularized tensor decomposition
- A randomized sequential procedure to determine the number of factors
- Large dimensional latent factor modeling with missing observations and applications to causal inference
Cited In (1)
This page was built for publication: Matrix Factor Analysis: From Least Squares to Iterative Projection
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6150367)